var/home/core/zuul-output/0000755000175000017500000000000015144424235014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015144437601015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000306066615144437411020273 0ustar corecore ?ikubelet.log]o[=r+BrEZƐȒ!ɦ[M cSy-Hgf1pg @~Vgiζ߷;U/;?Dެxfޮn߾n^ՠC4%Ϳf\ϘקbzuV6/?;|Yݿ|y+vŊ7 >=*EbqXgnxh{nۯSa3WկD*%(Ϗ_϶^ +SI211zysw߹l;] Hs %yqf2=;OO pzM.v=F|;F|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kzn!#(-aD~J7FP7M$7iXύ^$26lDt#3{f!f6;R.!$5 J:1*S%V!_F([FbDY娍ԹiY03`Cfw":ɴ@=zN{f}\{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}+׼ s'ϨF&%3KpNGIrND}2SRC ss]QzH.ad!rJBe`V +|i}}THW{y|*BP3m3A- ZPm/OE'c/z"BRr\lwCzDiQJxPsL]3,=M`fͷ~Vܭ54>d,Qw+C}F][UVYE QGn0Ƞѻ>.ww}(o./WY<͉#O5 H G+o6C9zg|~ €G.y+Vܯr KH^>z<h\d9:bġ7 -Pwȩl;M@f̞Qj_P]2@vN (C9zO|$Uvވ+ZoIfzᑇy ^t }|#qKrdK\D2s&[#bE(mV9ىN嫋{o嚲W9ȝQEkT/*BR =z*.h4(^&-Wg̫b]OB/EFδW~V 9;Zp0S]UIĀ')4 B^R4t; *퇄up}du@69آs;DPsiz39HoN λC?; H^-¸Z( +"@@%'0MttwB~R2EL9j7e\(Uё$׿atyХ*Wt1z\+`E8rVQUxMҔ&ۈtq2Q'Qgjy"}͹C,a4A{C U =T+o!er] g"tta[I!;c%7$V<[+*J:AI \:-rR b B"~?4 W4B3lLRD|@sfځ9' ? j럚Sʼn>^u w`?ԖQʡCOx]*9W C;6)OSCOkIL*򰆔l=q VJީw!pN7,/M(.ΰdƛޜP16$o tBFKskޒ1,%$BվCh,M%<>"ہ)Za@N}YJz{ћrۉthxY/$Zøu32EʉD'MS2}t i:Y`cФIX0$+`قSᔙD'Ad [kPF =ttSE *b9ģ7$Ϸ Mo{C Ӹ!uWȳ)'jw&+uߖt*:͵UMQrN@fYDtEYZb4-UCr[٪L.2teB ˛"ո{uM ]}EӇF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}+|]eO,v^ȹ [=zX[xꚯNI7c<ۃ'B? x1(5AKRVF2ɌУլ F "vuhc=JS\kkZAY`R"Hr1]-oR[^oI]${&齙4q6y{&G`%$8Tt ȥwj   %*e5-wFp,a~fjqq6xY,d$`!qIv꜒O"T[1!I!NwL}\|u.b3XR\(L _nJB/_xY6# ſԸf}9V}']6C9C7sn/e552s4IFk^ك󨔖I@[ tWv Fiw9J֥WmN^<.eܢMρ'JÖŢո%^gQ=p2YaI"&ư%wm+`VLJsC>?5rk{-cS`y_B}V v,{*1ߎ% qƦat:=G=vNv dߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B=aK`3CmF1K>*Mk{_'֜dF${T-n,'}6ȴ .#Sqד9m5zoX#Z{/ҭ٫3@/%hJKZ|Q;|m쇲= T u)1 QLLj`K -D,(7N*,< JDA?VǞ©H\@mO~W-ce{0d8o X]ԏ޻(*exBaEW :bT:>%:ò6PT:”QVay @ek~iQ){A Ă۲D+'h=TԫeVިO? )-1 8/%\hC(:=4< ,RmDRWfRoUJy ŗ-ܲ(4k%ה_' c9*%WyΈ W\Of[=?+ednU$YD',jߎW&7DXǜߍG`DbE#0Y4&|޻xѷ\;_Z^s=M\&+1gWo'Y;l>V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~w w?38v?Lsb s "NDr3\{J KP/ߢ/emPW֦?>Y5p&nr0:9%Ws$Wc0FS=>Qp:!DE5^9-0 R2ڲ]ew۵jI\'iħ1 {\FPG"$$ {+!˨?EP' =@~edF \r!٤ã_e=P1W3c +A)9V ]rVmeK\4 8'*MTox6[qn2XwK\^-ޖA2U]E_Dm5^"d*MQǜq؈f+C/tfRxeKboc5IvsK TV}uu{k s" &ﱏҞO/ont~]5\ʅSHwӍq6Ung'!! e#@\YV,4&`-6 E=߶EYE=P?~݆]Ōv ton5 lvǫV*k*5]^RFlj]R#Uz |wmTeM su:s@ -Mn3䦴mHЭj !'T9Xsl o:d lzzMvYź ^ ٲAPm쪊m\9htwmjQ\c5&,|^C.SS P󂏛o n8Fkbs/&a[s~W &/.9tTo]r8-X{TMYtt =0AMUk}G9^UA,;Tt,"Dxl DfA\w; &`Ͱ٢N^8 n`т ti6{b?-X;|im̓'!n&.TU n$%rIwP(fwnv :Nb=X~ax`;Vw}wvRS1q!z989ep 5wݫK]0/k<'dzM2dk–fl:[a>֋&". څZg`9r| 5G;`.4&XkĴp 6l0Cз5O[{B-bC\/`m(9Ay@/ڠ --i!M5mjozEƨ||Yt,=d#uЇ > l]չoݴmqV".lCqBѷ /![auPmpnEjus]2{2#b'$?T3{k>h+@]*pp桸]%nĴFԨlu |VXnq#r:kg_Q1,MNi˰ 7-`VCpᇽmpM+tWuk0 q } 5 ¶]fXEj@5JcU_b@JS`wYmJ gEk2'0/> unKs^C6B WEt7Mu Iヵvyne 0= g`_w\|8Fj}GP 9::3(6e™nvOσ =/Dff^f!8:/p6>TV*P,rq<-mOK[[ߢm=ȑt^, tJbظ&Pg%㢒\QS܁v8ac#sĢB\PIPfwJQJ;Qxm &OBf\ZA$Ba-z|@-I @x70 晪MV)m8[6-Te@`E^0{ P*27ެT|A_mnZ7sDbyT'77J6:ѩ> EKud^5+mn(fnc.^xt4gD638L"!}LpIn?j2ɘCGOa9C1L PU:LNTPlI&N:oճM\Qe%*?vQ~W  yr3-2+=Щp!k2wu_~c9'\ॣwx"k%oTͯ܈'i1Jh`(D"y@ "0#7=OP^b5K 0Bt&n2hev/nw OZ鋋rTG_7:0@Iuʙ?&Ԕ8e,žLG"1lͧQѶGM]}yxZl 0i&],cj/PGLU>u8.H#',c@V8 iRX j~p_u)oQCCn~8'hYǬSiYf)<(j%\IG.`ʌ=}㻮";?ޮ}Qrb=OI$(_F:lb6sJ $ oa"$d#HXu`\mzޮ63=nUu u~xUF2Q~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊$AZLk;3qUlWU Ry==qٻҦ62L0ډ"ܺ_z9JNȯ=@oUI y4z%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!o\戔-QB EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)@kF;'ta%*xU㔸,A|@WJfVP6`ڼ3qY.[U BTR0u$$hG$0NpF]\ݗe$?# #:001w<{{B\rhGg JGIެE.:zYrY{*2lVǻXEB6;5NE#eb3aīNLd&@yz\?))H;h\ߍ5S&(w9Z,K44|<#EkqTkOtW]﮶f=.*LD6%#-tңx%>MZ'0-bB$ !)6@I<#`L8턻r\Kuz*]}%b<$$^LJ<\HGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dXV0e[g#B4x╙✑3'-i{SEȢbK6}{Ⱥi!ma0o xI0&" 9cT)0ߢ5ڦ==!LgdJΆmΉO]T"DĊKٙ@qP,i Nl:6'5R.j,&tK*iOFsk6[E__0pw=͠qj@o5iX0v\fk= ;H J/,t%Rwó^;n1z"8 P޿[V!ye]VZRԾ|“qNpѓVZD2"VN-m2do9 'H*IM}J ZaG%qn*WE^k1v3ڣjm7>ƽl' ,Τ9)%@ wl42iG.y3bBA{pR A ?IEY ?|-nz#}~f ‰dŷ=ɀ,m7VyIwGHέ 2tޞߛM{FL\#a s.3\}*=#uL#]  GE|FKi3&,ۓxmF͉lG$mN$!;ߑl5O$}D~5| 01 S?tq6cl]M[I5'ոfiҞ:Z YՑ"jyKWk^dd@U_a4/vvV qHMI{+']1m]<$*YP7g# s!8!ߐ>'4k7/KwΦθW'?~>x0_>9Hhs%y{#iUI[Gzďx7OnuKRv'm;/~n-KI`5-'YݦD-!+Y򼤙&m^YAKC˴vҢ]+X`iDf?U7_nMBLϸY&0Ro6Qžl+nݷ" 㬙g|ӱFB@qNx^eCSW3\ZSA !c/!b"'9k I S2=bgj쯏W?=`}H0--VV#YmKW^[?R$+ +cU )?wW@!j-gw2ŝl1!iaI%~`{Tռl>~,?5D K\gd(ZH8@x~5w.4\h(`dc)}1Kqi4~'p!;_V>&M!s}FDͳ֧0O*Vr/tdQu!4YhdqT nXeb|Ivż7>! &ĊL:}3*8&6f5 %>~R݄}WgѨ@OĹCtWai4AY!XH _pw騋[b[%/d>. !Df~;)(Oy )r#.<]]i-*ػ-f24qlT1  jL>1qY|\䛧\|r>Ch}Ϊ=jnk?p ^C8"M#Eޑ-5@f,|Ά(Շ*(W-n# rve{椱I |p)U݋7yJw&PzDgi xs  xh\L r Ѥo Zt(I >|$>tnMdэoV#ہll/ؽnA(ȱbAj>C9O n6HNe">0]8@*0)QsUN8t^N+mXU q2EDö0^R) hCt{d}ܜFnԴ.2w⠪R/r| w,?VMqܙ7;qpUۚ5Tnj ۝jlN$q:w$U>tL)NC*<` `)ĉJآS2 z]gQ)Bی:D`W&jDk\7XD&?Y\9ȢG:${1`+i n8=%Ml%İȖb7AޗuV3A7ำqE*\qb'YpuHƩҬV nm=Ɂ-2=|5ʹ zi ' ׹U>8bK0%V\ t!Lku`+]c0h&)IVC)p| QUA:]XL/2La[Xѓ F;/-rtx-rei0hE˝ݸDt#{I} `v;jUvK S x1Q2XU&6k&lE"} Q\E)+u>.,SzbQ!g:l0r5aI`"Ǒm O\B!,ZDbjKM%q%Em(>Hm 2z=Eh^&hBk X%t>g:Y #)#vǷOV't d1 =_SEp+%L1OUaY쎹aZNnDZ6fV{r&ȑ|X!|i*FJT+gj׾,$'qg%HWc\4@'@—>9V*E :lw)e6;KK{s`>3X: P/%d1ؑHͦ4;W\hx锎vgqcU!}xF^jc5?7Ua,X nʬ^Cv'A$ƝKA`d;_/EZ~'*"ȜH*Duƽ˳bKg^raͭ̍*tPu*9bJ_ ;3It+v;3O'CX}k:U{⧘pvzz0V Y3'Dco\:^dnJF7a)AH v_§gbȩ<+S%EasUNfB7™:%GY \LXg3۾4\.?}f kj· dM[CaVۿ$XD'QǛU>UݸoRR?x^TE.1߬VwխmLaF݄",Uy%ífz,/o/Z^]ݖF\\UR7򱺹...m/~q[ /7n!7xB[)9nI [GۿsH\ow!>66}եl?|i [%۾s& Z&el-ɬeb.E)բA l1O,dE>-KjLOgeΏe|Bf".ax)֒t0E)J\8ʁ,Gulʂ+lh)6tqd!eó5d ¢ku|M"kP-&ђ5h ^pN0[|B>+q"/[ڲ&6!%<@fpѻKQ31pxFP>TU?!$VQ`Rc1wM "U8V15> =҆#xɮ}U`۸ہt=|X!~Pu(UeS@%Nb:.SZ1d!~\<}LY aBRJ@ѥuȑz.# 3tl7 ]وb Xnݔ[TN1|ttc‡-5=VrPhE0Ǐ}Wd|\aD;(;Ha.]1-{s1`HbKV$n}Z+sz'ʀ*E%N3o2c06JZW?V g>ed\)g.C]pj|4逜*@ nBID f"!!*7kS4޷V+8弔*A19`RI/Hй qPq3TY'퀜+/Ĥ'cp2\1: 0mtH,.7>\hSؗ΀ѩ آSNEYdEcaLF&"FhQ|![gIK v~,Jc%+8[dI368fp*CDrc3k.2WM:UbX[cO;R`RA]d+w!e rr솜[/V`+@;Τ`5d0ϕ_Lع`C"cK>JG.}Ε00e>& 2䯫vNj31c$ i '2Sn-51Y}rE~b>|Ď6Oj~ebIapul9| 3QtUqSCxTD7U9/nq.JYCtuc nrCtVDƖϧ;INOKx%'t+sFUJq:ǫf!NRT1D(3.8Q;І?O+JL0SU%jfˬ1lމZ|VA/.ȍȱh M-r ~[0AG꠭y*8D*-Rz_z{/S[*"꫒?`a;N6uilLn<Yllmb rY״͆jqTI!j.Pٱh s!:W_´KxA|Hk1nE6=W|$O -{]1Ak$ ѫQ6Plp;3F$RveL l5`:~@c>q,7}VE-Q8W70up˳ A¦g/OEU:غA>?=CۣPqȅlW11/$f*0@б 2Dݘrt +qrx!8 J&[V =͋A,z`S,J|L/vrʑ=}IhM4fG(Ȋ1{TT%41Oa'$ 4cj{ƕ_1~l"R Mӗݜw(u#Z8;dIPmj<3gLjM%1HcTw=(}U\4K@ ,UI2ߕkU .n".<"}URDIe;,ʤwOr ƾ*J%\H[V1j^MY՛NWq{\~)ϣB6Id_M|ުA !*ib7QJQXJl; }g5(JJό}fu{g3A"g[ XD$MY|b(n$-߳<O3ٌy g]iz2Ro" Jak2!OgA E#.ei`Ka`=Z"-;'f(AyL" "/@D$0kgT<*'B6)OLL 1 JXa{}]X=wzA2 1YuQHլe5kG RoI^W7&,'$U*L-E+k )ujDM~0Ϻ0 ^FV%z M'$iXL.AL|7u%n N՝_iZW =5_P =O2> -#trOy\8=b<%IŃT]+c0 OT$d[8<2єRy8:QTqwz5;o *,/HiYIЦ(!@$ր]H|s2q{+oTS4X;B`8/x΁#3ljq-UiJd0qbdW< OB802p6ru9^x/]xX/"zeo]("5G@GM-3Zj =V+^'JPXbk /FZE* .1/3A򷏎$_[ݮ+O%UDyE_à GeSuG3jc2O`WׄmGw[QӇ8ۧaZr^g'rr %Ts?F% 2g"@,m@MqZHB%y."clDXPX6'bwfpbH>p$>y;g;cJɼDglg{Q9ա 9:'[%PPC3W*o ,$e4/ ffoO1\|i}4T`ڞJu\b"1;sXqE>BS9ݽ6BaD@P1 (xx&Ԧle([YDY}nǺ Ϡ79ϒnka G(K+E>Ww"Pɛvȷx51-p L3fx@0E\1R>5ya(R:ji*Dd4ycu G5tsd u@9q((~_ȺcF.i%edm"N׃ȍ+,<}Bs$ChڪQRn[]&8ږ\Z:1N|2h px#5A hxj TRByvҡ%O ;{_nkYRb $#GӬT?FEvWE"LGG0-ƘҬAZ ! ec^ڪ9OZ4 P[yc#$co `l9sX$Q/`.d NK42H8WwZU#dwۇF&p1i=qt/"RGѬ Q`6 ̖y4ݨj_ˠm(8 5FrM-UQ29zoKSٗ+<ƴ ɺZYg?: D &T2'4ڝvu֤ Fطuu!ZmMa\iS-6}s\ꉪ}=U OcX/pk ̣nk@<{@ʣ;z۬{W\+Q55oo+R˷Z(h(Xu D6RMDcCxu$Um4ebqR"Xgkٻ~4B2zKN 2OOTm9 If#۽ql ڣnm_p(tjv5El70] w:AS4ۭj'Sqn}2h[KRE)S1W6 U>c˸:==Ϩ'HrJ v2kk2ڑćb.ͲI)"^MuRfF *1^aDu6yz iukhF r&XҩipuX,Zp€ͨSG8őO}'VzOb ]G~czv~kB$B/p%ޯf%6n![gV%֫$\P!ڠ#tnQoYmΥVG 7p$m-ɾCnpaCjѡp!&}<Ԯ__L$:RjsKR8b#ʟ 2Hi$yXr7.m>ѽ>O q.U=xwIѭ.u.bOlou eQh꯳ﳮ˴TG6ݘ֫LM )mT" eVQ䮻2MTE#$(] X|tV98{Ѕ&y(8^0w$umFF]ޙEt455ݾA_HIhYΎa;jzv ##=&ߪy4\NIh7nG]ɏ/ڭ:߭r])eKojbC}Dz z>Ę=Ϭ]iF<^#jgՎK3^73f"|N2}#=i;すB$f2{:3[Ԥ@.4Af&͸-HVssf"|2Tr-:̝kK01zwr^>.@G7c:;(٥~yBm|zҥo]Gw]vz;*Ҁr]hpGl,ݙx>ȝV{%:wP]0Pƞ McA:^4ZF&Ć^*G] oȒ+ oOz<1ĉ[`IٜZR5߷(MGb HvWWu]`qGؐɛ'ٴԸ1͗ݻam ]a X'W^P(Q֕ϙCS^0Ipa\汑`+ dJ…_+ӽغbjӾ4DxUm-u =r3c`=\Ng# HFo)n|p}^k$,QT>=5l (yeхg*T**zr s|Q^G'4a7󴼦v`V>q]zG|Vs6M`N}c(}€"W:ķUϞizIv<]Ѭ誢٢ 8@ `e=pFGq$LIk/k'=Drimz;I.o,<Ǔ?$߻3_uvP<36nlX{!Y</4`A"`1lUq/y(./v (`n$(:|$k!hOҭ,gE-9d{ %&E[W29'' Vǖ *{~W8:,Jn۪޽W V AЕ,)X'ҋIa"IK u"E }) ,dL`.]R^0rHC[Jrr9j&ޕ10D t R't d& ;?ؔrN ~xOiߞtxOu ,ڋkZZp;Oo eP,݋+Pz8DiniAh5+Ѳ:0bo7,[z|ˠ$t;Iz_ͻ#N3ȬZ@ ^An @ 8h/ο"htҁ1k3)A̦RmX@(Ma\ sg]? \DsJ "w"v#(H$wsϋ(1ֹ@ g;vO 7de9 WQ>gA v0UXq܎5y:p+}'\(tMr.o4y5JdKVȔlXaGLí8R2VPL^$&x^ou5bNgSnsP7tyHw ǃފ$11ef̦P/2~_PQ Ui.Ȇ95.(bQhۓNCs7As +vY_V_|BD񜈠(pUiaZ_p ibfG_;7T=q-`!b$^7 X4oڜhW[s}gK(ILGUt`zf&*`i*n@v$&H\PUZu T(!${x "lGA2~8HҫU9&\SmJEN}*aáNlMp?H;~\|6!}-$Y#}N~rNP|`ExYR2ɦ9T|T!D=HMzM(0pygT|Fzypg1|@W֝JkU;6Jԅ(ǟ]L* tE3XJU?,c8;j\h Z"S8s;o4凯%pGpa_/o 'Kt7j(tp I9fPq}O|3yt4 o=T6 2Xw_]ϮQW\ }"=Fmk-~a \Sd{.[5AOqFLH7‹8ɩZp amrvF347 b@i!$]?rmq(i¢#4UjFz[/f ifE nHG9mκNr DNr40t4{tAZ?\Z#/t l BwaH4p' hH@}5*Rkw8l!u„|gpvԻھnw貑pL'҂|{=rgSTU$sMc X^|# ( &)_ yG÷PN(ۂP=y'ʶ ʷ oO(-넊-*Gx"b BuB-u'y u ]'݂Pw{B>Pw BuB-'{ _'߂P{B?PBWt)1t p+s2e:e/BQ5 8f|:;|4Q:pr0C] r\bքEHm}JPN:7.æVUMHnEӧ6.~:Nk{h/"8_yHɻlD}:(bĂ 1LJHsޅ8F+&Fdqϗlt B3&m ,CEmO޳P?6O ~$ϳ"M8fh7暸ʮCحCuGUF4D)ijuZ3 T'piw C݄G`P_& i  5Ex偉Uԛ3 &#tML̒ʼS34cp6{&yܻ QeGK#up9:.bUq׶E,B#%A X(כTQs8!WyS6⫟%MѤ}I([U(Y3QD\khװb-<_1/g")n!,G՝& N/TFJbA͈;Y0'zj&,_o8ZhlF&SbJ=sS] \Qj8 S b(#Hj\9ȭNAUje/{pYc b}hV8}8_d$J qHo!],5rVs<݃6<ZgRAb ?Heў4$# WgMd+_At T?@>NNS fo;Hǧ>&^Aٟ($U>7 ^۳M^5Oߓ|`l72='RrϨ:ohѽa0݇y?ON+ "pQxgQyȗts(]?"NRH+$X.LpA#d)+Q6"T0Wy*?+guy!\.aL6C  d;O3D-XJ6$ZdtT h{`4΍΁G8".2#8`N2^Gy2fDLfXDs3HA34e$N[i;,hq%2jśzVr\$=95fT^)@r],W 1h*Wr8y+MIynUZt Z*y @Gp:?La_n%A+f꞊jRj4W<.ҖjVTy0WRjAMt;񢡋$חn;8=TVr# TO8҈ڌL, pcSlǣ+}aD#WO⾾>uY2V# Չ-0v[$ F tp{y&|B N 6TCfѢ='hrw΄ { 1R6Rl6k:4W,WWZ/͆$y:_iՙ0B5l= φPd}7x=:}QwI.B}X읪ZB`69x˧zP!z9 !& zRwy8r{hST`Ꝧy!Wo_'5IԵ7'H +YQ߫<4_, g&j!|$ X݌yPvõoNI|2Btģ'DHŖ' rr7A )sO[%oKjq:/T[%(QS)E)25(k*E >Yrd:Z]8B>2p`_%O7Nj?877/cOX'6 Q"匳ȞnTj5A}FiL'XI0SQ"$\-_ r0I͹xgZwCp'ӒWb$&9Vk> !d#?a'p~#nVD6hXӐz`+KUYflA<(A1ӗ[16$\| h- q2%Db,Icb̜0<,vJݸ.%x]YvA3]g!8T2G 8'tD D4Cq\H,ֵuebpY(ĘaI0{6> P0C)!^31\@.ZeILi@p[W8G4_m2.0xĴ mߚTbq1.l3 v_6gR‘d`(k' G ?>Ipc3KYC#[QZ*=D59X#*o GaY0a٘`PP0fȃ-YS GWr ;0PrFa̰a2t/kv$Y.҄ٱ\,y9S\!j)tq >ύG=6:e+)%u`B.K 1fCxiTYkҝalZ >KE΢p疳M8狱Ğ4G|cGbQX]9B(g^Ca .0!uX48#1DXbmv*HL+{bրy`FB6#+{ؚy&j3UfM:MazhlR9nfՕn>nbt|As^&ͬYl1UK bacA 2Ȯ7ZB2Cc&Apt焂B^lL[Y3L(px5"bЯ}j&]d}Q[q^, \&O:0ߢn)3kWQ = y8YN5|~izf\fqREi=+S/&̗ lEcֺ^x^'fs鬺?mkgQNa馳.E?Wn v!q5U + mh.^hqkh,Wn֛Ӽ쟗`JjNpy2obd }I'h&yNA%%Y79B@f GR/ab>ٿ~0 :j^5ƱE^Sj(W/ 2OsH˱s ҄TP<2Wq1.i:oIp]zsvr"@pp(=X.t =i*{ȶ$L FtOA7!dtcb =8bɯ>Vvp;}vʼn/J>T'`DNuF %JԬ4,S% a\-5/EXs'3XAA!k^*kr5X΍ Ib =q(p(aPnuKcNHqU //7֐8 >ӺNrxm6ڛ 9-GKiyk$8v^!{ɴ%M8̚Vzo0 {=( UWq"v#rp*ujǔ, iߛ\س+Ebn0 z8<+at)U3=G-Ew~zGBhoK& NY-@%v<-CM Ђ9W8G)QʊKQx&Q< /_n_H Zʫ|9|K'GQI* YuU9'-IW헇I.11|xѾ1 ,%2FVUgHN%0l Ifl4+35O zV/$8ڱظ`hV2L"5υ@~^W4WLf16 aN}z0Rh{ۭ|۞oHj5VQ A셁ˣA0Cd:^2 1h\h]J9j0BkVѸ׫U)*1o/7&IvK'{͊6^oq03KeV!JSsicyC `̰y_˱$X\jd-uAP/,Xl"g $5t9KLr--ZBbB.5%##[c1^f)pVp ^J2i9{Fk-{1EuXx.t5fY"Wg[.#kKN* 멅_aӾ~+ K:ia@ch,8J% c9G*ڛI`+pT|R.YAvPËm֕%eʝw! "tC_^$%kp=O: !o0#})?]z=)]bJ[Zt)8$4!%yIc&X2 _<[ 6$K o)pt5TUyx%I?!ls'Ar0fry[]KOʙƜ &,-:Pz[)m(/8fY^b@6 1(wU]3efL-]!X~Uu9)˨k>0?Bh_v '.?|\S(gWP~|&Q^y/x59:j* -(}]w!P ,eP-c5qmb-_vHp.S/r *srmԟ#LJ֣N8 8BҎe$[O8d*%ty'nuHIKpwm[v-y'¿V:\Ʊ``eE^ڢjXo[(W<Qg&͓`¥ C1ʌƩHɬR΍U|`wFa,]$#>[IN^Ĕy2B]x{#Mz{L1V{1BātXVZz_<gHq_tuC3xҞS2!h4oߗy0y# OivOy#SQ%Í {k1. {CnFIBke!F5 ,]GrC*35 %jĎi+Uyq1-I\g!^ ̐ڷq_j2͏C,VgsʐEkXft;Jq&,~*C7lI#%zkxޓoH+SH[KT@*SZG#fW%;sƈľusHpt|[BGS HPY1oa6=S| +(P8^Q3Zʶ"JViK'ݞ81ňѣSUBTqӌH+MkJ̸QFnGz?M=2egX^r*gڛ۶Whf{]Nz$mA0^KJI|{RGd3[I&1}b;LP 0IKh78FJffsZ>5X%Fd&-8:izuFʆbxZ18: S,0al<ŵbMG +y-.eqf㬸tIo\L|Wc[K V~)ֿEo}_r>*#Cøؽk5aEo=h:~IvI*k4xSy{eo,0u?ʀkQU| C˓_ BT&>N/=8efj9<'  ETI?\EC>@t\QF$`O핷^70yzGf<э&U3<8'/MO1TK󳗍W;_Ԅg0^io.Q:'fh 꿮(R< YjY)ȍ7Dr> ~N1 G]=FG%Ǯ uX,22Ӭܵ)ҝJW N g8Lխ~J\ϦSsxm2r6iks c4e#^:;z=.ѕE?޻s3 ݅C0{·mlWwMON1( 'w^WTAS7̊YM/(@{r6?>1Pz ʁ` ʼnW}UIa6f@cOuA) $iD18NJWx#?=xy76+0\<4~M\I^1ڏff/v#w}jtV 㢑3QU}dyr|2 yD5;&KA, 셷z5Q]| 5=dn׍'{ 0kUͻG,_[= 6Xlܙ$g,tds{ކ? 7Ë__,lKSJ J{[]G.)UէeE TmQȷ󪺲 =yUзp%5F6nKGI`kJ->h.0^_-VK/"XR8.O{'L~2ⓊN6Eyq#7oYRqL kr·خzoqoӟ'NչoZa֊ISsY'K8hQ; ڻk>ѝ6#!gN w/_a,NMeU¥K OqQȪ .iK, qß-o^Ydϻ^f|}x& !Z7Jϧi,YgT\j{?i5K E#ē?n>o`AG]x˩|]Z3&o"h'Zk-B3ZB#pV#Z;0ZbkϦ ]Rc%U1QDuB]B L.0qpr5mB][5=9)Cb!";"B  Lw!vSPo7^^²S<.Y 6ҝHI>޸SNBo9iB|,TC! ;ݰN#D4hnF+h .rdŬ8>fa;_RVGlS?i'(RTۓr2bT4* `[yvL'-Xta'a(3 @ū?_\W/.HKWA(xΨTû_~E_<= peu夘}=I*05"DeS".' & zl]={;\װoyBzpXLp J,z@MXgm엤qͭ9:L7W $1R&D$QNhgULuV+%QYӴkA"VU-BSW!]}~an<ͺ}ܼ6:K#Qa1aziX}4| 8$^({tM~A~lutymԔ?~V3إۑKtkvM@[dVm:wBS:lod&^@ߝ-Vi砗6I fR]1.b~a?gS3}(;<9(iOxhա}an9jmfסxS$\&Zc?U)vω]|Xv&vyNYm<P#2{e̤ا7mqĿ_`閿HΫp_w*o[_C<-PwsBvWkq??q(Ѿ}>Vps)|t bGKHY;l?@>HF}{$fHG3~_~d;; Otms ;oiBj@>(R3X;m} 4's6(g}&@o"5 }'_w,G4ghQk8c]HS &C`C Év4Q'X4%*dI8nM;jM;/bں'ær͛b ֈٺY˺9rgC02 RJ9@$rS.57FX()756LpA{v胮ɉ?78c w^6~?>|ry8)ʧo $鴘:{lsЗ(rZ@NⷳH wC %8qz饼raxH'A].q+a!Z8Yڞ>QJPIjfA=mi>,7tt(;,kxefp+Kal~wʛ-)A`f>ҋ)-x8עlv񵤔Ul7ݡoagm]LYX\G!`>^1+뽀oϼʈ0m(85\OBJQ8b=q,E-@9 3啃1(+3"Abp:@4?88 !}"#AykG.P+<ïGL>E&^O1tn /͎2̇Q3rT ?7?GW?=XiXB6VQյE+q/{ 92_~.::C%e܍8x:zxA8ڎ|q Yͪ ]>USo{c_pcdQh x6:_t@;,\?HjoXp)G`W QHսcaGe|v޺Wcd2> *̓V\g*Q wXdޖ*KJ$ySbPm39 YMl:H3[N.K MʍUL=)>3c !˜fi%g-Nx̩cn S%/V\W "!%(wdLE LȠQYJHY0 fkP6l@/HWPo <>l' bن.Ǐ!<wJ5{/(lK$dW>~頭wގ頭ݧ6E OE!'?մ6iRS-K+>F  ʥT+y_ZqRԶcO\Uh 7e|mӲbgVTP܆J8>^/KLTd}AU('*0thn˭0+:W ߈ojymb@bj%u~|. ~>}/ PQX3csc0ҩƔ d6;bׅ)& g@a͌sZ; rC"3n*-рRR5r`҆>KJiJ['*5Z: Ҙلct;*iPF&-p mgvB'$a:I97 *H&H4y)qCET1xS3'w_m 5&J,+z /4CJEpt}R}p" %Cj)IJp-8sq7J9Cu!MYgf/<>Y.MD1 H;?M;UC1Db/nתHH1g S 8UTш@ c x$s!(N`31Mԟ_O2:@"JwJKii(PwRE*nXtcnڂZ{k+VmJWѠDM0tn7;ni"iba)9O aݏ&XٞLU6nE/@ꎥXւ&mmn3yj Z%\pbMJ!Y@ȅ 6ݨ(x~j@r:QQw<=c)yaF1~a*p}W}'[ b#*Br%XcpY0ks VZjcR.uJQW!!I-8ȊD q i¹z0 Д9j1ܡ D&IF Xĉ(crHgj|۸0+C@+4hR<"XYrS~]I^ْJ^mίxAڏÙߐ!"QgFeE✶W34jatgUt;`ЌgjF| )B=m>G.Ar63Ya'if*j`D璃CM~2LBHTKijQD6ִ0 .zm>-c}kO8OlӊX(ļBf6l', J 1q3a9[j>׬'%ƶ1APl_K'ƥ潱m1_i9_OɐFadϴ#xvB#ZK q vF ٔ|$L1\(alQ ʩb`1%0pN%LQ۪1k$CČWeqNF 0|Zs'a[ P60mI^iREDCb4ÊtTc`>5 cpoYc1p)@DɯG~QAƏ !Cʆ bw86o7'u KJ@2H{f$^ B D090FB%#-wv:n KLq=E\?ذhQ%fk&^p|8@vGdO[Dq,X?}sAi-"G?0hC=޸_i}񭶐aا%IQ6+7o\+uxC+8o YX}ۃOioAG|40Gq5_0dTb4~p2|َh7g0XcLx$108w|@$\b7 D b,f{Pvը,PDRᯗ$_?yR,..l]ݰyG9vigB=W"ޔLLQM! 5!Dx`{\8JB&h;>цh 5p}h@z8DT#FO?Z"塼н+Q s/10 I8Ţh%9~H0'Hath?I.O,5oL}9p(,qR1D'xU؉;x'QiE`RI3qU .8hX80u ʀxYĊ ۭ5+S2=ayV\1a fb#Ua([qj(a\TO#?v}eqa+ʛw }9 _&> gO'l6ݛ;8KE}h*K&|?qgB ?l,SxSI"fvd\6l`gx1l4=)_]k|8wxjyM^ï QD۝~_U#)|hԞugwsÃ#ޮO`齙[^Bajk]AJ.-Ef E%g;r /R5`$!@ +8ڿIrDomBР-z*P{Svl7~+u v !eC62vW>hֺ2 bf C.zrigIZrW/ON.^N]q+<ȏp'wϩ)(_q۝ݰ@ cY~A7 Opfq}''Yh'&a8'z' nI |+jYG k/eXm}R4KYCQo)U4GvIK}.{ R`Z; D 4+rOXjrryZI|f4,fb}f_ndK޻Sf|zQ.4O~٦-T-T])\.ey~5_Jo,04#{IʼnfKXbBO6b_q7IZҐw,5<ۖO \寓ϣ(}yDݥ+ WVsN'ÿz!ҏJ,9'@#w14b(N;vŽ07Aûf\ .n4.*-'ҿ[WѺ?f_Ӫ|C\<^riϣk Чai`p] <hET Zci%`Iy#I2fV*33*L@,΁qc]3SS}`FqG@VF d "7,44ϣ?ƠSL a)vu09` ߑHLb Mr.r|6h>h *A)ܷ1 Sv]o튄K]A+<ĞalS 5tohuXa}3pP uG0T"q5x$\ֱlV٣,`5G)ѵ=c\r Ư\) r;y4B2ɻ$rʚGN70 ݖRu?W`˥a_kLs WNgy>Ejn6o{YKmS?󭏃Xد[]JT2O5nd>M|d5jBj9 L0,Kq ͟%4[ 49tjo-2/q4)e?V3Pa5 Sd yռ'&/txGk۠xi.*iS~mGڼﶢ7>1|mdiI/˵O\6)>./_|%LԳgIXjOVnz[] _f\Z.yWyyˊԬ~M*IL$?}TrdaA!{CTrz,(B5 5T{`̒w{ $6Ҝx2\%gU5adUP Y &uRJi2ES61^y<zjD֒ȟU>s78)H4)N2H&jʍL"i29ONOZ+駂rsjW\Mָˆ˃AW ۋ6)FIU:#׊vȅxJS {{['֞]hD}adQ~>EKIG{14yWN<0JfiMoȌWo)Q|2͓--@B4\|'2`8@~%\2bjJb%DKVݸ3ͪELw_gggaye1r?Mp߆?7߼? ?=0~l:q-Mwͯ/瑩e:r7n- wx{o\An1^~緲[46Lmiba6_@?FKiYGEKRɯ [5Hlply[K ig$Fxޢ:1?3h :u:4$<Å* bqf U,bSi9ͼdf3@Qg͕pgX:OLu{J0o^=! h͵n |vU$yyys_ t1ved<- 0oOw$|\9xdˤ!ap!F7JeVT#'ZЭM/78\՘՝Kbgw.3KʬbrZ8Kh]~NJhp -W wŖPKFCwmc/1*‹\ss{l:\mw CJՒ3yf<ό<3gxF|Gу1/홎9`RSJ74(1H=3F(,%}:X >G"I S"!&G@q2̙w0l,y:-JH86'W۹2lݶ&JwW߲L+cc>eh/}H×lIp}Po>}End~#(CFÅbEbTSmb 4ʤY8TOq/)&ϱ֧bh3a{V8ƏPo[PV1iLwkx]-}ȟUNvRg]{oG*Tp[eK61~\SBR~~3hDqFd ā~TWWz7T <$7d+IG ۟ =Δ>gJ=>XX`3{:%"xbAsX {w[aHLa]3c8;:nTlk~@b㔱],p+eTqnD-ެOYPB`.? Aj79bnBi|tҶB*Z0N)# T;jA =K5rQǎh9}1f9I& 뙱1$˷`&q"Os)C'꜂ 4D0ۂ9uѴ]%#_?#'}pw7St ? !A~H|C ?$ȃ C ?$ C!A~PC|oSbt1[x [}7.;%aA93蠩#δ0!4ػ`s;m``"VڼbIc' з&:x‰![*.me}KFlCF"ԙ,e22NZN̲/U,;լf.WܲoMm}B_U-Kj_:RYbdga.~x^]\ިT DXh勽>>vMMn6wHWJjTI7dUl1Mj2 {* ,x^7q5ڎ)."hitֿDY-/ρѫoǗ=9|.9ggޓY,pm:h9=?'t fz'3毶dcXh0& %Q1.M?+8.qjwk;7Xqt6b#P9\IuOU 7|FN(z"Ϊ*Bf1ۨ~&gQQQT_m=?SWe~wkj\m@4dl|nYϛzO۫>o ~^KE `z #Kt#oQ8JJ8k[ŏY^Z|s Y|w?ނցLvWrV}qc6#Eo\ћIJ‥ʂðP J)iܯJh( N4$vl3hnP}8ESh n%"cB嘐f5Ս@;Yzs1e`( YS`KXQϝR VoxquuS|;M&nX@i ޺FT:?}P:4]gP PrP1誫3bk9b0={bЇ|aaaaSj`h9< 0@aaaaaaa'@b9> 5T*δBP~PWBBJӇCtP:}(~pM tu~qGob0hj'hj.UǗPLf1 h w`3_3a2wYnϪ9{qUh@`}!9 N|&|pԪ:m!V9PL+ZK9i KpY%:q|=u2ɛdB~YO"H<cΩCrcpX=}/G f.gcgwq/GSv`cc6禗m<寄Qejֿ>f mmY*_KWմ:Td eKP`r,%1Zj6;Ag,w1+×yjCm6z~uދk Gd &{jdY2Y BQMVWʪbkGʚs\sF^-櫚j79猎` |ygcZq&3:j>.>dl_H>xcŇكa(=Jd%Cak%كJdT,7}fxYSYȅNKMΟgTlVB-`Yy6k[]@IFs H8!MBĹE}<_MIՈ)Ly&E.s6g%SHb9F5kҠ[o\ue=٬:d_iY.&p!W%! ؇1Ssc9c6Z,/#rcX>j씖%S2 *D I"C7-2~%DN7A72as89tĩDȈ\d cRK%M]FޓL9k43v2w}[3NA w }S|rH{Pyͪ4^""tzrNaJ(qRqH kgx"ZPGbyyO+NvNUJtxADЇIH Q A@աzV n"l6pGg^MWuOVZ)(#Efm">ZMGRsd" 9^|ڶrWwoհTC%2";LJfv/eB+h}d$'.&gJ0]i+"]w/2+]?^`uxH /_劃Y8RUk8 Kh/>J2~.'MT%(Vmw_'|k"' dEϩ!'CDdNG;:%=y8)3"gJZ0$$΀0/kVg˅dQ\$ OV F$}*nxܒ[-#\Î?0:$00F !ގ[]Ѩ6ڪ/ˍ:UG(KD!!,`=8Bwjd"~lS2VyrB,:~4r>>EEb Ġ,07Mp XF-v@y_]Ϯ:;SD4H۔3|JIa_óo𬈼/?-'x')=EJi`TX(F ',1>tc0w^w] W9agt2HkWrvE]k)2H3fu.WdΏ"øuzΥ,n`U0뜨u.@jrSIŧfM^.".=5a<ũə:"zxOG߳$+* ?ފ>T-Y"+*^чt5zJmVk宻k?j[= Nt..,_qёњ|@6kN;前Ka<.;ngV)֛e+"d%$%xI Ű^bI*–zO[$S*йXKE]xkOE[š6G_uOHĐ-ҞɊ{8Wt艱j]uG~hG'"oac+VRIV)ӫeuUHb$h, Æ"+B޳6n$W ~i vY,p\n˒"vf&Ň,.J#J8cXլkC)e)x(b*'@[r57ʓr=˰yVfco eVJ\$yLk:qc/&Eg3o푺ܒ th਼ T3ėېS& ibI<'u:~R;)5?>%1u¤95$Θq_߆Fql~$ _LQ` 8MpTarNC[LŀT}Qmf$ggV,9ԴhA]~ M . LQSoVDP1]E엧dPU h\n`y&K>ڙo%'<,~v,;7dN:axGlB]<}7+bM.Ƒ>w]wItIN<}bʒи #:~w r7&bp)l,=,1~kh)~$ N#,k_;siK4']=fLI9Y&tj Xe^͊uq.eQLhqV.7\$Ԙs LK$Yʊ'cV椮(]Pj~S)9a% SG.(¦GhhgOLs`GF;n 6V+?2qҙ\,spJO]VGV1:m]b oU>9'zM桳eilXǩ!*op1 g1 (cY[{*#:V\FXnzDsؑ*5^hmhGQ~d f s0)l?&G CC9RRt)'SkD1Z)0Αž^?6x8ospFc8 4Ӊs|+WofB"wrnC,|JGu,hrTj ~94Kpz5^tJWj:k|E z}p0`pIx | 4U-U8Diʧ*=zg-gwc.@OUc`=sLA&T%{Y kӑ|1%Dr<~Rs\*Y*Dݏ~sY`}GhQ,_J%759\F+gJdz~}.,/)__oD~M]2mL}S}0s'fM-6NXC߿ k&8YP([/7狇wk(/F -]W.'|nnʏwKowý̷Ei.w8.>ywnD`YˉiVFuFWnk׭bplQ[X\x;}YVb_ͼJ~~<.fEJA\A+E\̒n6]@OѷTg'LuCׄWYE9MUjIpW[Z%ZO$b>5K}]b\F|-`l~andMvTLx.7e*@=CZ͍_Ü~S b!pGe~+7L`jÁgz/38+IcIA{0G('jE>(9ٕ) o+?Ow$X;,dxk]Qh Zx+*ev5TV/<o6!DTœ`&LD,& Dxf6'ej Uog6|XQwl:u"Oq~57AⰍs=3W]uHҽ$T@cW)[us  z &z'ڪ\f 6ryUtr}W0`Z0Y?`iPD:u$e&U`Oۘy_SRlmA_<P9au8 IWylm%0 nLtw髠.E髆qөj?{]A)B4gDkΈM$*:/$#FQ I/ˎ >tȀ^J[BjyV\>a M 7*;`/U0ː:Gp_ J4D'o'I-=Ԝw^yVnf,'8eY6.&x-n9M3[F l ̳Ҡ /o}RreZa@# 5Y3ɽ' 5k`4(]ȍ1tZ);3rmVL ZMot2#z7 tFtPC=tdBN ~t<-A90[ϑ/,Oo;K|ױmoƒF:Fa7Sb51;W B`*[SnRn!c`=LT$ l=TK+=rDZv'nFKnLPc%Z O)ûkTГgY樔9IZ5tw"u{'+z琨qMW~atC`n 倘dv#R";Q /QH$ʣU&%aBN' o{u"uީ,l0Skf=Hd4xD0<۞5CćvZ| xfSna86t;j2JDXC7FW̮qr3>5;Ě^8)ce&\O5tsG`DaGb[K`y,LĆe46#{/MdP2T%DbcOY O5}E/!tK[C8`J& y`*+c/BlsȠ:2g":k8h!fʹ08E Ò%il%g["vO ;g8WiVǬ N mГq+JN/)ޢP޾,<ڢO|rjLhe ED0гloL3A(Ss9gA¢jGx>n²t=c8 /i=kC~ +S%v-]&f&9Q23)?l{`n uc۶N$ T90s"$9T&RؚmX8#1Y'[ČѰr,N_Ox: h(cYtj{\=Yh4,~$9kUP6Fl=ʋ?ОiWnjZDli`)e`>ǻ""NEgГ?sdD)g՛sF%p=w.g& l|8@e9G(cY~Bg>NDC0&_ FgPo$lgDĞþ=펋qQDz\.ƆFZqaLx^n#>Ey։UZ&!3X$DZ*yDRχ=똧 X6Rq74?X ,{lp$i3}djj%!@g.*Ur^ҭp!ʹǕj,'`:SZUxbX݁cz:dv!/q3KNj\m-'sua^Z"J \B0gQd`j8f~<7dã Sg譖 O<"GoST@қig$ ]9o qY,Oh .K8qi8LBHE 8Gx~Jr=묳 :fS|W-ɈYMb2{QfyATEf:'y-(RG5>rJ%R[ n"+) B|dBUYČlgVY< 7 se;`e'@4!eުzΈмRIAkk 1Y[C]}\<AA)8y0Hs<\1E]JÕoy bZpvC۷ | ugx,^W(t5I%BNt:F:e2 OQ)<#ai*\f2={ b1 | Q5g+儒l1h{ >c{MN15Q3AZ̕ן=ta O01mD$7O&?N=8j}cqIk& 1TFWDو)ԛ$]\8͗G'pΌΌG' "F&#,ҧam= _Z>d~Zgϸ̶1;9_d^.zE?uOw AЇrk|rk`lh)G5Bx6EUˆ޸J%3{<w9-Hc&9drb\ey3)fh8ԵlvO!A˟0Ȩ=&$2xW@0[ڠ9MEU튪($e<-VrY#wr$d$|\{PКx_5tteLjxVTL;Ve[bV dS9_n/pȣ,oE2~z#Fx'=R6Z#^IGqB{Q?L 4[h`u[:[2/THE;t7vXGykS^j[a\ I3/S%6D 'BwIa;Z6>oX5t>xN竚@&cV)r}_e?"ZXQk2(Nqj j>cẵ8 *8ʀZҘOC`^vHXU.|ׇbRw~One;S$ |DI)pAOoaS+RCG7_^Q9zG[OsjEYa/ %łNd"X 'Hz?1A~8$Y9O&+E&ɒ1 a@9t83G8$B(sWZ0PP pgJ yl33p&TRKtJ=UzD (`JZ/cTʈ6/$˵K<ĆxJqVC8ۉH =bhۀB|w a̯zu+>ƮK1U:Ѯ<}_%tch>:S *%cC#J]фuEӂx9w̜c߉2&y'Z:(<اlA?tŃfiƽf\d>nCO'3Vf0>gGU꺖@?MxXQ~wâ( e}5ݯֻr>>|HvmA&)Y,vwEo_ɗ.>y^?6גeiJ 94'$3!H[. p][|_vo灗jUP嗍o]_Fq*t]|O+->&]0X_r/nWu]/4rE('L6 ;|ٛJO̅׌Epچ)WfX7n^]ˆhJ[2iF_|{,ڿg%\3KX'am@|7ն q>z:XFEK*k_Yjmɓ<Ya%Ss]jTr/H{Z-P,bZ>^=/˃k(z^wuy: ^p￸]ԁ/{;\s0/ϫƾ7z\_Vm!Z?nZdqc]}RAV_%mD'4qpbRc~/"Om{UgW>4|ϫBp&Uq.g-x|C?5a.v:bv` U}k1 d^"Q0&iA Vǻr ĿaA./7چ2:rϰP,7>0?j|{t݅-lsc8%+? 5Gf#XzjdP6 p.m*$> V)U:VS(eRb<\Kb %{׉ΔKRX\V 8du{ġxxK񡨶#>nl@ɗ1I5J6nV]{_w[V|O.]P;ʹbFhH[kxv bu#?Ma?%dڊ`4Z8SI<{{kgJRHfp[faܞ>QEz^-r@r-\9)* vmF@ ~%3sQ~_M' )Rbh%z4(%5[ľ*N>m_k> ]ydoa\Rb$.^o$)xj g./ֺ"ϸt8T3%[k ˨6A\Tĥ ɢNQG!oH":M}~0)әsӒ+4ێRW Y92$E툃 C.gt"G MUrB *9%X;˕|h1 ].uW;ZHG Ŝk#fuX}l:=DF:F<>K-Bë^OJ^v'"OLED`{1{T\{Xr!/vQ|vMߗ .oMsƕ ~2h^Ƅ$!j&^'OT)&{l ( ܡuX.Ezo&U=cD4UUA8INY>sqp O2LyǑa=J%lOMVTJ5 #)b'afnSBm*øe( Q5-PeF6N,PG+Z,ϗjugjtnvW }L䚅 dMJH (b:8%*lĊA=R y_ʂy_>޼w:i k3gD=#BBV $LٜdIN9cQF:~a"|*YNYX/8B=k:(a`\8#Eq#q?1 euyC$(w㏁RF`K@}C9\[e\=OJ7MsprR 54鈻0XFƮ E.X&F+8w4< +qqdA(#YfI>DXʼn.8Wdק;Mj.ߚN ̄FYغD i0t}zS3rTOع2qȊ^}ooZ *LHA@| X;iXDva5 8_"Z_IxƅPU]5/;wW(ecgpOƹz.fΗM!=~)0i˜:<QJl÷V:sFOc NIֺ~#41arGܩ0 5O.4â}iI AhxG0S%r$>c"Üq;N29Jnުqy[(88Fczxb8܏]UVx(ELNU $qM;.zrp:(ӄ_Q`_ DBQ?0wP q,"f'$6*wPeT@% s_ņr9_P1 43$̀rL.By/<2{޴7f?XK9[1Kfn>}" pmy!J<:`(*KAhڤTFIX B/q2;KCvwe; W&0n 8£Ѫ$GųrkCoKTc$/a\ILQT݄}pǛѶuQ -GQmq $|s'c /u\rxx[ /NR2SP%7EXd\M4L[1O5v-0t lo덄Dƃ{{-Er$냀OIJч\ۺc!ޖCW{ƶLcb]؈.n]Ʊ҅.4~"{zwzW`NA[rЬ>c!'l+ez1txha"wm{*A QjV 8lUeA>4߈"ŝ(4b8^LMwj& Y>ѧKۺc)ޖӥ%L( 7^Vd^ͶDPmKL!ѿyQ,j5ӯ֗6fi<|<_BM1v'ZIKZI{/l|G% x=LJqMST"oDL!HOva+]l_XQS 3m`woһ n[E,i~[ ,2í H֥nDuxkA @:~)y=d}˛_ws0<_/VUou˨j5WOwq S;} _lZvoFC/ N=ջzջ})*|3Cxuee LZop׫j`~mnVkN4w[ ?\_GUoWt}ൃf=氅 0W=wEFUYAW XJؤ|Gg;{6Aqjk\r&@Be)/U~g˪/o$0 f,5jL)Pus.fxŹ7Kj)D qy{5;'ʋi2oLo¾B8$Im,KyاeT'5xMFA)#XSLmNP\mfiApػ6,4&\A:ll6d`3^% )_ ϫn--V_@bK,UWxS9i/KmRq(?<( d ?H=$G6 1/BC(QmG6s]0, VA[R Jɹ r/hQq8:lWz@B?hoayv8k(D(dAiza1aRZ|X.m~&.욕([ջ֡Rí+BeaaoӬBu?]kLP-T0ɳ^)|K)5{ڕ@CZvx;@j ; 4}bJG萎jtq([;SnK{.хKe#143r\j 6RNu1Q =G!GІiF%oeKȣ \pr&*JlKD*|x`^wd!ң fʐ5=dt6pt$elhcdHr.BP2'(KEf3/v]%9r( -erPHۖYwG7)RgY. yrS"ra5X]zt9ta)] @vm.o_M8yBOvAxK@cyGktg|tow%iog хN4f+ޘhH*[hRR[ u6!@;PѥF_vG3^y)fuQP eU"/RX딴LFF@<x͢Α/o7{S}g_ZzPiH0v65@ kE:.5-LO@e[j Y.odl:finW&ZoAZ *wR( A堓n=8@vf# YNn5Ā?9sWC1"-YQpxrR6d JF낃,l&X~zaFPֹo6-iXS nL0сϜ9n&n>!*8 NBof"\ R 7?gQctЌ9lu Lv!+nbRYM_R(3AP3Ή83}B9ES)ӆu?RjeN(093q%-Mx#%AUvwDPp}n62WDy @>Q}5ݔWSTxwݭ٢djŶiu]J/[ N]U7(E$bKriEv8)*?ܲKrrw{g ݍ,;vmI&ȶjqBG J2[e(ꝏ+e<Gۯ̉II0meb$I/`k U cJ5a jd^ wJx->c=1d8kŤ^u5~{Vώ徻6wuxhZ8z İaİ<`ۤ.g<|׌. -)V1S|CK]A r`c\po޼ ҹV:F ꄽVG‡c@&&VQ6U$y3f4j|5]zŢS?a=rU!ot; xc>W? ݌w>⠃xW_kJ/VG ' RFbR1`DfQQ,LY>cy.Y|#zwy] ǐU-Bv~]8ēm_]wA[#Ӷuk] x86}G,;=5[Pwva]8Tur`rh>(FMˤFu67W^^]("&08~Օd& ጔzָ#C/[ѵ$5>B֖}pΑ}* #*0AY g~>Y~6/~\Ʊ^8֋zX/q lz0Qs3UxN^{nwxwY`+)y|)#ɼ]6w9tNvF ?on,1L4Ks? t3g-y 9Zr;3-MNyhc,K\ι`"vvsYNi54 sk Z+sN]_Gsw?}6ovxƁr<@ןZA.c^/7pcf2ׂejܹNV $u(Ey6 3c066l)U ٢>*ڂl)Dx ȶs#!$k2쩮Kt Q9JP'(b̳EhuT1C1טgƒIF:{=%s[x%4;"T NS.^oYO("fކ)0ꌙr2ϲ Y_t5NEj_fqs,W+C2E5U' ^Exo>]~U"Y8+? 4ܛaa%4 ??̳/:W]-jtpk!Ǝ& 6&%*/;G Z𡛗PL7D v~qZ~D| t>ul5ym\v]g,` cO__d_~=Y%cI#*|wߞMΖyNΣwrwٳ TM[g'4|`ʡBK=z=9kZa*"2k?zTC_H 8vR-BpJ[H )Ea0UD/s"xMO֬-OF}4fgËt_]AKdh^$ \i orNdtxCi4&P_p@T#c @=jSܜ-^,\Z"]_^S?o}]$b<XZ5徣Ho=~u>&SvV'8,uj]N"6@H*ڕkNͷBKT{מ=ݣ\BFKx FrJx#y r 8MF.*\B0A[R JH!- p!fpt>w3d13sP83@Z[af;q& B䔶NiNiN6SھSھSھSھP  )))))))V3[+ gXm 6W^\"  W3j#?gA61eHgX7݁1=#ljWC6q.gGaɴY[ WFl_&. 딦P- a8-7^YׅwRG(VR0N(\O=q,Pg$$0!S&ApZ]dW6Rʃp-c I/5.F5#wg=l |j Ј@fpv~{/(`^*")pAO 'm?єDK5R)< 7IU@TCIDcK B"8t nI+/`8cBVo%9Usna\1w^w.#&gLpTM.-,g_=}Mӫyg<\y$~5DT\}U`jg_nnc"uqBz7f&X55 P+([}+ _J7ws9Qpa}UO$7{ ƮqΩR]J%/nK*9Zi'C,YN٬Q~^ʼE$y 7h\&yT"W,h{ h~c\zQY2Oc9շTvɰXlfu25 j0(b7$hb+Er7a :#s3Rϖ+awN[lB\:XfgwQc]9u UXdn_dV|﯍eTQE=|5ƔGEY˟*;ܫŷb:_6>3di1Hˋ߷yJz)r|GUW_31f>;ו1qϪi DrcnLwS6&W)2\ {YD68d+Irڙ:)Ύ2t| [ɸ>j$yY໽] 3TW`u`yØ Ve5NGG]xi>(zc7@ ",+$4p!-)X>(vl7hn>EMaM.P >u'aofm'Czh38!3=,' $R3p,xlD8f RVͭDaaLxԚL ca/y~9bט,!ʌW,1 + [<)[uM7;1DBk1·تdSI@} -GɓʍlSh?/U iB<xv0Vݽ3?`-Hg픤ˋ n1}_6V6XHtbeh nK)SQ Ƒ6ugsfFNI4E8SM?cq4c_x[eAþEy[Qz;qhXYWpVAw|[܏fC2$xan/Gan/%[ s{an@HKE /ЅB^ ]x /takX /@|ڲ,tl>Sۂ`*$II;גU5MRtF|~c@\sKI-%KIr^J2̥$s  P$1A|.|.5KR|.5K@RT.EKQRT:D<07 o> $~dfjebin\-Dy[ qsV Yi??_VS]ʻwtGiE/|J(7DFe_ΤVE!5XxZZ34`)鄡6wo/~jPJ<G׋L&?>iǙ| d9 \ckFh xM7 XjnHZm͓{1M4[k34juI#F}`hBj58e1z;KHm%:n,e['qi`RI-ߚw ZGKVԳDm՞:#,Ri‚Sїvn=gpyl}f͆aRbRv,A\ۥٴ&lI/;܁"[zDx[ ;H5GR,xyrOTcX:,8G 03Oѭ5A h46Zvd {bu@#IZnդoBϩb/{Κe[lO*d&:jr $t NZ'+ca>@ ChN{^u;"E RpL*"maaQiƝV2̆UAP4@2Yt8\? BtOۈ?8VTg>"aHD"TƒHDǷ}Wx7+RP(t\,)rjA4DTU I+$eiY*#|;9Sr$ 16,-~hh~4E$y t ί H)P(5}z(и%LegDLVڽTwwRŤ`^5q_mn2m,'%CSW*'"gS KGha]RL݅õ^']_g N[lB\:X(!0]L~Mv"wxXƃMk uI&o_*J7NS3ИPZ*@$!+t#d_m'qưX儁_"ÔDjtnRv6"IJCJS[)+V00S|f^<jM&x@hkg10*x~9~ט,ʌWv8ikmŽA ߡ$V&X׳;k>o)ՂJLL;^5RR-Z;>Z*^DާˋH1P4j=vhu* Ҙ9O/-C@ .ZR;7 #$V6^Ԗi"'^q<-ֺݫugz^:- 3^#pYWҋ4Athh !R3@'|J4Q:đ;jFL."*>k@CU#2z0y.Ť, :R>B6xјI&N)ռ81{^kK1EaAqË N;^Sإ4v'C pؕcgc & ;&gk@[(`s0OUo^)ݹ[{ &cՓѰFm?%֑t ZKj1\+fxx4g-!:r_:=J(J":K"v܏wvb+*nRRIo;(OѧmW#glvP5 ֖S JӖI 5g̡ YPC+LD !xhXHǞ"_")qJ7D"q8C9Ja89Kk8ׅ#o|Q1nI4g0K2?|ؤCcoѣA-L|[Q*mAH[2F),B0WY)mEyg화6&t;bȆq1*u~tmoNhLTͺu{1[U+^LD~ȠE-zyG)Pʍ"VFudXDcZ)"V?D h$m>gvdc#i,\6-4rT_.{uLgzq_)iӲu$@`Y`g6ڷq9N//u\.;U>'|SyH)QȏEV)1Q@ '(*7jN-P}ybÃ|_.ůtZkw\˰;-L0[3J͇ʑ_W S>.[p27OQ[eVGkpdHtgg_N=4 K)(?{ =ίBKks RB(Ll ?Hv»et6/y;<`Ϋm4@&/(k `mgZcbVXk >X HmXIc-78Xkw*ֺnZ[.aהj؅7teHHhpxV?:l /sE+ǧɤLc˰wtIF]2.p[=8٥D -h_yɴ yٞd*,h?|,Ջ[Ow_ٜM.y:j>;#]T_0&x5elqDIXG5-UP@ҊX0;hW*yK\m| ,άR6.d j'UdUJ](kC-&`oG+5~{emcvGwB|=>ې96ǝL6hrHxt*F4 ,gJ0 JP $m$Q1 Bq^Gx`}a,} PxKX)C@!WL UY&#T"]< *nbmҍ<rlHr2{q/br zgBp3A)w5W:kLzS)X^Spi= +;g嫮vj^$+-),sf}n̨5{6GqmzIOo齯1! @ RpšjsJ^'mzVhc2w5^Kwj_q;oYA>vy5drw9*$->SDVh M'(5&+VR5.']޳HQH`omb8cƻD}~\(AV/9o)*sEdchZg0敶Wزm &jv,o}AV%1eFn^GV/E4}x5TKJ;;y#m{ީ Yy2U+Vth+VQF*QV 9d0PeU:8{#\ݸma1nOы*_ RcitieH-OpR :Knxwpg?O<\AaH :w_~&֬qY\!7t8$Vf@o)^$㇣oiكnjR~iREa=]]cH?T~}9qgDly>jJ+deW/8pI24.͟x"һ(kz^Bj1tU0IyLİ_o]dPg/}~<5!9(:H.BGW1 {{Uih45>fg, Ï&QGeP2d|uw"}c,|8Vf˳ZumͲtK;+)«j..͟fsFպp2*x+CX0҇9;-[~Q?iMJy $N)/&~JV0%hCZu݆3$÷~f=@4W L+\> G[,jm0aBZ:sTE}Av;=Qq*ȒtDO5Y>y^UL$ \>8., )yȃ'@%Kktɡ!|Q{_Do}1`(b#dūjSHRIeRaEc/h=^{㑯=mAeQT׭4n^l;.ˮ!fox]q_a&5r?wţ/pY~ 2340{YD~ۑ**l`ջ B;l/³_ەO&'<(0t7%l1;(LP$TKn!CD c.[Fsf=s3v\7$ T̪5#fU0vj\_f+lHBa1f܎*ŋ$(*!JPa~[TwKMcu@"fK"qq ]ǵ!փ'/Tu +,cAWrrL'P#c3YSJ,?ڒ)Z^38{kz2B؋s2 j:KKY =G70T@rVDPzU0?BI9ݛ씪MXRZȞB {y-̀UؚYWD`EK,2 ga){3dFka9*f:4 =WVZƴ }@0+G =hz#^\dɨYi+r#2 =6f>E2hQdukm"%5GBɱ2 ='#{0VH#*hhºTWB {$ Ŷn+!1hG W`r+Q#q b,a#n 3ݙOTf$Ҙں M5%ǘЃymKmi&UDnI`L6hǘЃyVJx5JPZiUc{~{+o5rYJ l+l?BQu}T"p'/ *5)z0OGKqS%ST((IA8% ټK5Y}&1 ;C>Bb0q:r)8)[/-B9g0`_K9/Uf[/@Z!{o`ټڠCiՂ*@; 6Z։g/P~|Ͽ V}<Vyqx3>yިsoW„*1 Uv1>V*):9Zbtv\Rޕ)6oͻ̺x% 9~2xF7P vf*V! 瑳)0oka@䪤PڱuZ5B<+BZ^&^7z%( 1?BPow@(ŗ׭Vc;tٚHs .H-/?Ǭ^B[I(y -eXmmd(3vrd&PXNXc0Ҳl^!kt!ECD(MV> y;.Y B*öƗ\)03-E-T99vz0[}dY^'P " Nz$?By2{3>UD5`KU ,^Z ]#{#d÷$B*۷P@dǘ@QoTDVl2XhY]Igw3?B=bO.3'{PxV-*ȞB$_R E췂PFlO@F@ 6uA*-VzW@Z?5"S(t`Ql< -v\TJΥ@cOЃy ^TAq *hR>΁}c#n =wvg1h{s]w=Ik}%Kgپ%j%ﮥ[|vC{N/xtPS#CQEp#c;UP%;)d$ !, lTqWb uL^a]#ʂ@+hS\vUqUPu<j w&`SbY:b8i;.#VPut͓ {N 9NXvDuU(0y$[!A@JSZt ,o=;*t`! \؂7yNʌVI^l{Yc#F'ǗFQ]9xgE^_q!Gui8f"F'2G,V[vDU1 d08c{ç^fpz$8/fkѬz)s&I9DŽRn&x92J#"( -J)#"%k#4T7 I7]s@'E:?'9g۱AhEF[}bnЧu׷nj#vKnvv]3l2 ܧ0r>{?5^y;ɂ6l4Lj8rdzXB_ t̐b+}o3fovc7'YkB Ӫ̜z I͚Yˆmvp>zo{yw,g'yY?Gк&*W+ ՞:-I0F}`hB2t8e16 &CT:Ӎ Gׯ/1vrdaR/1jTY+I5=*b1h#P*EdRT٤7ϳׯO9hUkW:KaL\2h5Z`(3*h2:ڀmT*! j@ C/1W.):T.~w'7yxK&%=2Nu!_cˤCIRo^Ɵ*v)3DSkL"TqF1uS}V!N__#%=2}Ccr,m?LDe>c k >L!¦V0#(Mΰ g%,#6~}:KimY.mϝ]<]swz+kYl[Lx1ozm?]'<_~wo?a/$i0p7jET3!xZ`rF Xh=VE=1QbK`~rF_8Ԧ;),aG@ErSeQwAkDZX?WDJ#RԘ੔Yu*5P,DV\ k`N9-"$ Xi0fQ,Z,jw8I83=0}uK Q9n 5d(p2FQ#Z덛I[rFJ;#UAhtNzVu%Ul. hP ܗhA_: )XL~LR0 Mߏ/ ܿ_ ~7 <N>1Owx LM§O=I"Ō&h$P(⛙t/;/*CP Śƪ]x|t]2jkD9~[JsN3":!mJxMqb13N"[%>%P.7 9?iTbΆw\5.6GSwG5;0Qь3&O8c4Sv1*۵ *`ُ_5ǯx(̫"w'.O܂-|\_@}ҥyeM<(ŵpCzգJ@ yWD7ǝ2d}0q Qu6rOf5v Y a"ԜY~5{BgfKvJOaVjD?'z-+qӞrJX]k4=MӌW&y vUzUsRF#u4]lUV|^p R ?sqO{NCZU+DizYW֠Q w&TY[%~%ND\6 ֬3R'$;6SG\h0O5f'БktK{2Sb<K:! ԯоnfDg(;}f-W'ݠ<lW#5pKLt˪QfM"D5_o^W.xzXR|٬ ՠ̐7 \rضrwQJΰ̃0lsY~ǵ1]M6Nug틄ڞ Ѭv$(ĔF[Uu(+z'{k塗*Ǝ`! y=O .&a_J$~'5=hr}Xpwj5WÕ&BN<~wRw~5UL3Sd|WMPڽ_euu~22,s eObF)X_3ɬ`)$Ť;&a'vަr][΍6 '2z#q&1:OH@" "M M DC\T3gw5pu@gWFoh|ʅ Q Ps ](5±S 0gجσisslo Ņ|l]nxg:Cm](>f;45d)yVC^ks]^.EPI45^i'ڇKLh;}mgW((@ʯu.H'.z%L;;?uζK7Q/}mSʋg{ Se޴zsz%T8 ySJKI UDpC@K,s$#uM[YUm9ݝ X"WNmQ`LR#ymTy| ZV3K4UV1ϧ+ů*EJjubM7 "]ucI[A6w5aNZ'QZ r+)ҕlsyI$%4ϝ]>ӎζ >B3\L9.i4ol^yNjrpJ;cNdv m=EU`v uDI}}Vq}! E)oMr_^NQ`MnP7>T2n;.`8Լ_av оRRU$gmY*ǩ^9>at WCx@;lf%fKId5+bSUC5XV}NZLUܧ,pЅ*v=6_v8[Vt5RF딦P-zcDN#+:z'%qb%Znդ֫HMC؄ A2 ǨU1kQ T"OPzqyVچSt!t1:.Fަt][+x4rnK=Ohnr_1E+%~ahJwܣѣ8q?wơq]o$Z#+Xu8$Ꮭ6] ;gjV"FUJp-c): FF1' v LOܸ}(Nq T$_$c'ϓ)n/,)JQS-"GSJ5eä3 cB ֻ8;Z> BO:M,3tp'Sv<qIl^g↝9J?p!˫*] U!Iqnp/!RLݍãތ'=ބf N[lB\:LpQSl|~N$ ,>.xP#$Co9(ww[x'YLg005&nUӀ4-tM lZC(& ;QXXlwIECkFzNe<߱kr>p AHRVa SS1cp豉hj4BZ"jB޺W}ò~' Q{)xKm> ר]|0T"- MeS ˽)WoƮp2z;6&L7\ g'9j 0N0R~yy< AV=C(" hbD_i luKuvLj 0Q1E*$u8!6nIԴ}TUrK%bdF/cL.lמrk};HMRtlKYvbMMh/`0'N2*0D%4驐N0+D{ZDJGt`ZD"rR5!VNtZqXBg TPf $0 rtmHXG, iB,uxwIiK|~n[7Do LڜWpczHzAu!u[-].vRhQf#6Y3Uc&?tsA"]ZyfZ]~i1jr ˀ.&;ϫwVhܶaˇ8Br>ra<3ft_kULa;&\{W[s +_}ώZ ҜѺ;wm?2I..΋:LucyJ1GC7F _=J' ,7@05V4>iUxLJZrd!2??aA8#TS} ? ^,xZʒsG`dp&W)\+cρ +Or0'm"y-ִnj6iKhK.y+]8*z21ceKtpqcFcm@bzhX+)-ܨ uc1mq]OY~,UzGoZ0CQ206?AI[vjGem5(_}E_JvK{Q:=Ez`u#fFہ6nx'/3~M|IJST Nlj6rg)B5@xlgL*wۨ]hFzFzxjqM@H^F~wLɣnȵrC|ugi3$BJ-={RrevkjP;tOtrec=mUSn>`/ǣ_~G|X3~+#Tr@?_ [8f0ŸXyw-[hG3FO%Dkh*WѠUb{S٥I)V@}wМR4ūW깼0{;ErR.Tw.҅,9 SwV1J&+u(Eb}{MQl>Z lx% ,,5yE0!xk9741h0Xiu``p  G7*+.IZM1t_n6;LtaZ>1Ǵc^{Jf|lS _ŷ 7;PpH:_Wocv,<AUQ[ j!jΛx$orevlXrP Q5@RSbb(;k΅$p"pI&xN7 XsN9=)Y1`D{$E.j+y2s(Pʍ"VFudXDj@[hHH!doloԼfI- VAݟiY'Dcz8?3lGr¦ A&̚ ^ND'>0A5X +k*eq$RcÝIck\ߟx)aR/1",D"|M-a$EV 1& P?4=~zDsg ss5=/7h?{ov;i};y-Xhm 4wFm6*GT KIw`i7ʘ#'t TV,w;w3vI9eQDTe˅i\1{7&fvig)%i.@(&30daxŘKo?VW,Xe?#/TvS:Qoo>5CY*!,xr:\3 IDm)39e?jfp6G9Ϻ=h&($FE(BH%#ܪ1@Z!dH_H^|YqD3X  sU!I zVrrc}̶jf>xR0|`)hS\̋ly EFg*ѫHG0 Ke\w"4ҋ/sfaݬۙbd޽Jj&R&gWgHp>g X*kb$s>}7888s˭#:F ,KPFrMI,`rJ͙"ǜ M0&,l0Gr^U$G"iSlp8u4o߿*򷓷OzJN^$V5'0ÝfjDQ*2TbOƓ(8Xs8uc1RRO sHX:EU`2_ - ~^B^5 OJ UZX0^Z05Y)`VVWD@%Z\'*- ҃hIpu\C=Ճ`ej!+"@:2z>sJDEor*2ֳj/5͑n.Ղ.v<0o j HC"xl$mIs < HmW3C@:ܰD+wVWИql1S.dIK# wڎ"8>hJ#дe ˑVM(^I`%R4ÕW`ђT@4|4`vYA8mC~jJnO ٻ6,4eiC@:EL`&ޠN1E2lʶ}_u7)%)RrE{a%%Y>D,jy i, [,m@Aj-Wr~5"JcF1K)`)$$dP,GVOVuκ{DFo3@$F0iJp@T M DCuHG m8rv#UtNnoE$fw n @-tO_ow8ct;À;#pNщ>LC{`(;w=C78;g"WJ~.Ε9_x"׸A9]]%iⅫyއZo2}3p9;Wɜ v. v=ڜ Og; o͉kvӀxaΑadY7g7X2{&CDTN)RR(ݍ9 Q u8) ?[?[r d@`iլ#Oa۲}M ?ncR'}̗×5Μ33cf<cX֙C7UČ≙bkɚ'jZ^VLHz_ՅItעR}{00ّB%nݟ(Yʝ[qgIoۊ[+VGǑDRu[Mõo^vLy'ZyMi[*|%c~}ˊZuR.l s.9Ng5i -x%H?p@=c2cvS5dzv˃f}`=Ńi["@'}=}'7ӡۮxIg|pRLSNq飶ěU`̙KQ1$݆h:Jywrm (`04W,ﲢ |4 j-6B:`^e.DBYd-8,r!c0)fe  D9>0Gd餋=Kc/Ǵ nlKFbb$S, A9p0IX7fƻoB5#(a[o B\ dTRyiKFXyEc^wPznzvMr)9s05|{w<6Ǡd[B^GLFzC0FŽ9蝔juTN!srE؄ A2 ǨU1kQRk zqyqΖϨ]t륕] NLܦbw͏Hxe-BHxȜ;lx-K3P)gf6K HpaAU+VɝVZTa˪5Lή;Q"߆?ݛVi'(,~WgtaaTԬXͣ&c%SZ/>[XYћ^ǖs@a:¥}3m86$xT&1O0Vz>|" :V*Lѧj7 1LHqҠ$qK8i[kDLi}䔾T#m8f,| /S~0SUB= ػU"I*.s/a>)"ctq% SUn/)0eo%P0 dGtb քʥCe.W%|~oI$'pZ/FDDcSzDod//RtP%ZcTK-oXJ Bh Hq.YЫ}rtW_.]4-H6eNŸ73)(jЇ.)H@˫#fןAꀮ^&gՏ|6>bln K6|TWtNq(*X$40!-)H G3z{j3hnʚ})t $hÞb PvS5pK3IU*Rn,{;6o*!pP 3X k&9?H;<:3>6}#\T8iE{յ1be)b+NXAȢGRzk?-?NiрmP%pPZJL)0"4r!UbB:ð:Wv|m*:NL9̡Jx*sxT Č8[!@Yv V @&NFu#; ʃυ] }_?ujE>|aoc^B5eI];J@i+X^h0%^h# ! -ymKݩ뵻P2m<]ۑZ.bJپ9}S$ܧ,{ŒjCJH[2Z(?kI4|W2hٓ;;Yb75}xvf6s&|تSxK[֥5ʱI`hw'eoQx֨6.)ޯsӼ w.Nx|7VݳFnheٷtUi=wLMn*_`9iζ?oքitvgԗetYOd+A`,58b[.}f[Ƣc\WH{ ,d,NEҥVTC#EyeCeQ$G߫o._F@iI?T8i" `SSJN Uuq_*j(=YNn V}{Q =^P3V^ڽvfI-E3zoeqosovRO^f|槢I,T̤{+vm}f) 4!T;E~9gɛICdH`Vmdvȳ (-w=i&l"m/8^BRA]h[{La|H57I^'Kh/Z4YoS[M;~jt'6z9;te3l}?~k-----| 5VSݙs>suR%bHM-ox[jևp g v|Ǟnw69YΡ94Lq %0FtTTk@3N\:L3ppҦަKM`HEnCԝX-94wdy֏ grLΟ39&ϜAV5+1cW`Gq?*2߽֩.`r۫߈3Tzw9|ƛiДq2**3vyNm ESuQUn['}4 [~%d!H(K@i"Ş+Xf`fa@_GcU+aaiVA (lV) ^tp#``3xb^;y:>1D4F}`hBJ)5PKagZeڙӴ.h9!XxՈ ,/$R ~9Ba#JyIRyC&H:ԩ? @^FcY+냉KM-a$EVRD;u=ik\^||j`'"ol 0םвi|ݨ`[? \=]Ajk!({HiD,kV0 $Z:`1VYE8=`(3*h2:ڀmT Kb0XJBS>uspgkq9#$75֙l+{sys{g]=heпib5g.]&.^LJ9>` ;kqׯl'mF)-!4WJۥDCFHsJ%6h>jUւ"8Q`%Vy㻌.oBg Kҕ}pvTUiNrX.+`XOagT &,JhEYJ`M(+  Qu,"4X%fLtB4k $J"@=Ƙ9>OvcJꙄ نk]%=Mq M>FQr|}}0rQj}k:NMviԶBݥ4" OI 6z[@%PHқT%}Jgѧ1D2P_M\lr¸إ;qty-9޺dypױi};)=5"khy;i9[;o.U3ki ޻4].=߻5{R˃^鸩e,U'j -=ѶܡO5hɉS~Ԋ>~qٸ4sn\x_0tW)R|ٿiݭ(_t*^ԟڥbhe${QJaQ"*Yn.EUbEWRϪP e K 2>±̚(pZe2{ ݳKetRQNH=|w=zzrvyRCgVNR9zWߨ"dN5 a1DV>W!ΐHQs:[|X>_ql87BN2YJrX9JEh$ܒ⛰hip)reS,};#//Gdv^+^F2q╽p& ҉[c >/X#ˣoåZugC?J!-e`IY qE*ʘ`n~ۭꜳ *BfJ jp}J]^4>(}Ɠ =^!p"_ٯݞ nO[H}Ng-Űg !vSnM9.mG{Z_0K㛎oΒo tL(dת0R Zs&8S1䒜J!p.|| ߹lcw[8,ldD`5-6w}DV&52*e;lw3;tr8.OXS)6ˣV9O B.I$80h hh%$^-+ >Al2F '763sD&-@A[!xsH/?pEmy֣{@TV#J w(Y P :hU@i!ȝ c ,}`[9$%Z70xȉV-9!0"%@`MRCEQ`sHbÛ~>pe2./sAꝱ҃7\8m"cD#1/Yjcc2ZM#8 K5Hhfd:H˨\+@䦙H "iM3<5)2!uz0#c(41$$ w+m; JI}e"4$ Rz~IS?,,\]04u7p7M'u6gu/:RUd>ʤ/Ubhr{V~LbӟKIpR(iU/=7m|đ.'wgD*N%3(_?'g/3MMJɍYLOI}ckIIsEEhCS+I5[r.88({H3w67x =20"bnRBKor,+v2vcW=AoF]]ѯMW9y0EΤ%/Qfbp"IpֹGqA #v1"ZqNypu񢞠[>6D'Pv`=u]$VyA,vÅG\~uFJf9(،;@6r g:=rhZ4Ij#Our{< kijB]׫$ɩ nJ^Oy=Ln;4j8")?l8= ̨n~d9zf=ix obԗF% ~>'bF=wwRum,͊q%q |6 zP&'Go6iEq8#*y54<q,SVsRQZ I 4G7U>Q>"8e%x.6Pwn9fAxcc(9V.:kskYz===%=O< q;oI+fI5g.4(coŕWNd؄ܴĊLamNDž49i2.rk6>y='sKR\VetH#gSR!Ԑg"U!P~^om -重n{x\XYy?mP_O:nx8 0½g|SEdϽ 8M&B0{`p?qF>jf_HAH[]í.|v*|JLfA/.-QpJR>r SzvҚ VD[SR$FXBY9Ӟ(<]@}M-V _= >w|P&.{{`!>꽱aW-}5V{bZ?DWnIG(Yc"[?}^<.Hz *NIeK8lٖiw;ԣFqqw? `8s:[*!ys*yfPu3I2׊*s4)\(T@9'}r461Z b1٢Aj!rft=»w]4ȍC֨ͣF.nM}!eL`n]M[=[9|Qp.wiy_ԸDyK-7~COת4S3ucpnbz&ᎅKސlAϛHѮ7H[; ?4'l~y.7gsaɍ w(}|O0_4 ͰWSX՛7~{ x՛67@IUR۩Z`8WM:?C}53j<j-'t{yaSh`vGU]o#7W|Ŧmŧ ~ VF<3e^[lg`K-MbAZdwEVzt^.-УOZ~=^cq[7Ogɯ^}[Bc'R6-.̵(vWu>߇"v>a?3Dnsb?>ha;3MOx.pr,Upt^|(UYS!:[Eh号qbqN6D)F~=hvCx2x:gQ54-Π{~k W$g~E݀j݂ !&U脨$r`Ly 8.D\!\GɬwL0E, ~ٲR,FT"i' 8!$Ge2pc 8@pZ -$nK#f3s-ǹ6=\CE:[u!Ot|lNl5K{w =)6m6W~ %J%<2cʂ0^EcV:A6OR+Fxm*{ѕ\61D2U@.{%YrcPAԔm[n^NXHMEZ,EY)aΰ?a:@ i\:^i^YW)|f l j2 r NA WY$QI`rF*X+Pi%Qu&o#5ޔ06 כP="Xy`GK#^\FAlutkL]g[7b-R RYPZTqXO3EmS")yEΕM*!"JIwsO4>Wga,ll "!EQdpURɬP{dju @a)8%YfRQ*ZD ,Zhjm8u$D֖ MuIt__>u~/Gd}.! vDMiR}N4. |%a̕`aAHi:X<$ŷ^})9:G޾,6][L`0~e~ee! -p؃! hR(p@BZ Q(+a-bE2:0/]|@Oҽ3;@%&<T,8nJ1`uJNJ*e`}d`Y'yZ`J78j)$.i^]eq}k@i'zp\tϑP , zznV?auxyqi\TRecB~7iHF8;*+=NSʝ)!@qX3mC يdR6bJRʡLU>YJZ pѡ?RO4/R 09 H,r4:F땔9L$`Fp*=g#N>!m(5a^Td+q?Ի HPprPPִp&9myF_ܗ6Qd'\,9gQ.ƹS·MNz-A5,q%O^f\,_i>U }3M^gZnOb;`I=^y^GOo!]:'Ao(:k0uvm?<;oRvqfZHkkM3̒,㛓ndb&ey, ]4"V ⒒(+bQr Z/;Yez) ==\ְ]4{JM Lji|c܏/_~-5 8s gv1sŲ8t_:xaؚDCW{v*5GԶas 5,Amېqg>RrrISr+.NҭYݖq$TՓtg7niEtgcy{ˈfzwj6wf:'eʀf <0W?gܰ@VɛVwcJdb릋xFsuʜee6&x<B&k>MnGόb(%xxwQbC3N^VtnKDz݅)Ԟ:PN -L y%ׄ%WyǸ%1MW4}cW%ܿ@W@>X itd<ȄYAVr ƈ [DWt :?T צ`X􈷤|UZi#1@͂iD qM0"X,d੅CGOܕzwѶ{Z͑0: cHikNZ.Qf ΚR-q& pz|{,?Nk4ei2-$wtf݇%G & S>z bД@y*X+Pq#(U`5ªMPNp`t$fwD iX@ m 6BވNHOqsJrY2Vb ,O,[V<ENӦ(AfyTʓݜBRoa}&T? ױUZwVXo_WA8Gn\v|7ܔ5z?͔.u?L{?ԬX[TV+~a}Lw4z H\ KP/=MM).v$Ir>?Ib hpnѣԽ%@tq/en>T8%Ҩz!zmp"΃1tsO||]Y6E/qЛQ..zd$3djv1DRKW>%$;rpc`,yLiL`eW b~EN! ?,7B煢GبE[&ًT<ԅh5#NoYgifU4pK$9uSUqtW_,ݛW\Hd ?o fTW? E *`ST4< obz^(hyE?һ"Cc<͊mrVc@/9 Aћ4l4+G̃Aq p)I`B J$lX1cKAstS5M #D*)y*!.1.p5133"HomBARHͭxO*У5xbb ؞s>i+.kl.8c࡜Z}ӷh}e$pPyXi΄Wɩ4<7Ȥ+2$#hhX)z&eŶK2q|sbuiعPyB8ߟ|KWAm+[foq^~0>!f漖9F3r Q;ԁ+˘'M νka;I S|7Go~>>0N5ǣeXW9k l(s%SWdMZv;Ma9i7&:|P*Xt' 7n##O7|Jkrt&%LI,,rRŦw\L#:Bz(=㜶iSIt28W%"MK2ur&n5^'1]T%DDpML]HV E%đ(Xk7vE帩3UTU{ktw*[yk?= "SX"4ifv3SSs;_KVr{Ym59= I>ۘ )`y1EDXGb&oEwD~l;~vwuW\7zɀiaz Y (x/TqMkpd,(,Ot)`l2`|@ee oIrY@2L1E6hҵď31adobu옵P措k(IF#zmq̠DU%¨0O̵ƍ&xl?˦К.yX{.-SkT"p1+Kq,[*>z|ӫ?7ؔKZn͹n9yƬ`uof_-oؕ4'фTv.ⱚ=4|t<i=__pnm̕$K62Ak V9')YXZ /ZcyXO|ĤmWBLUI%,W }BE@ksYѐ8puNb_L\wO+YY%jvGS-m> r&誁/@hwʏntEà ^4zV&D_q䍫] kMTcr;= 7rbsѧIż 3 f%4gӟ57zn>b>_Q Xrkm3q޼mcSP#}H?~9ГYߕ q?d/yٲx@#D̃=T^i 97Khdn"wg mfe^"^b'7zAmf9WQ Bv O&HEgB)TvxzC1*ZT FUA,=-Xy䓥:>jh͕vRf5(XO2,x\E͠)树ndrVG,cJα=%L`N-`Bil8{J Y:s&L*JG"s1,} 骨 g$t$©ql922S!~'9q"eB3AL@B|#3=7^>=44?ĤʃNT Z {y ?[ <&Z  *. 2,^fQ8VXeA$B %]TUʒɗEJH2>3h~66=j0|F'iy:]OڽnOGݤSۻo٪ 8$'BL!06h֛D)u2ZqC6TpF9Ա,'KwrHg>@_gBI"bFzx*Br FitԩP:5 R'CA Z،FO,r(F_:' )L$@a$}Ms4ϙG2UtA(H6qf{ he?1nJ)\Va7v|&HS9=u'o"U}-1:tyn|zݰش7k=Ju[GpfZHk`#ƺUYe-ߴ|s|íDDeSfȱ)'Z+&Vh/KJ*BU(E()AeU\o WxGa݆慵GTMeR\.]˗[rle1er>3cLqbؚ_DMˊiI﫽P;EԳTjӰvpV{A6mqnLkķR܊ t|e8^hq{C=IcxթX2:#WN}.9Lg53Fдᑇ*t hj7ySòJ֝Q܏y=фLܨǞ:>ٴ,TZPeM}}MWnCsz'L@ye(j8yr4}|yl6~w ٽBT>BR多 k)0Ō0e T9f3UMy8&L|žQKpX}vp}N!CiM/.:ՊP"nCjPHKB<ƒsBrfBІʲ;aKcKuفr Ok*,FzNYHH1v=69<6Fy9D2y C <*e!cVD!Fi!Qh7W~NTW; ^R)gf*EH"EFmrKb*I <郻*.$PWaٜ5IXK *K&X}qUvtgKIlԅ2s&("/#"~άGltdY4%&v>GYz|?X RI# X>3MMk-O!LHC+>tj?Ib7oV>u{8qo +>]\ uƝ/%04L$;~Ҩ7o7&̯]oqpuqI8AP $M^θ!J$9*=++GqՀ]Z#%ʏF>FcRsW3,yE#GGR FFXfpṗ~/r"5Qpa2~yHlk/tx](zkUQ_f-V*Fh5%No\2WvU4pG$9qwط.y;ָf{4$dKaG^>\E_|! OvتF+B/ <'b&=]wtHꐮ:S{bUQ9[I,Mw [h=8A J&K0 !VJJ6, [vj3hn}])td(GɁ K0!3+ %/HJKҭRFM0cq@ºj/tkc`ʤK`ۡ 8O`qJK^{~ӧ+,$[P{QJ2bfʈ6g%Ѭ@Dc,ZBeFwHRljԆP&;$K(v)S|7goo?Wq}>|<}GUƑրa#iJx+e4V3![F;M~:v$ LO-@Gaz}xByf&UVirŸ{MΗݦS2: \pR  XxJ%.##eFs =`l[WU?By~NyL)?{Ƒ~y~?w'  lvcH.,ᐢ($E#%N_WW"'RДQxkJ ,5@1+m7"P) ϒ"8hNkDь+AkMLhX[ٮp5kKe;~Սkt*w[?=45!SX6SpLNjeie[3L׬͟^+Z>/7GRqF#-Kuc̡)Mc*j)=1ҡ01q@0H,:&%4DEw%4Rm[ֆ2nF)GIKmjjqN7!5:=BSsO~W`0$pyi"#{sȕJx%f6+1blb"sӨ Mk`˛^uvkDmhXJ; \! "pQ"0d$WPkCÞ/;t+1k(_KY|P`B@ ΄ б5ui!5Mkb _mv"9턓On5i0;dEt0I1LQb0 ^̉R},@S2l_~ly(?)~ûbcsє "SW,I{ѐu4z٦~g21hN8 Ė^z[&&3ɇ^oSս69+m"%ܗYUZSs] 2FJ5\R*AXKl'um.6jIh 3)XZ-EaNelv^łs|S_O_~?_)cu)K5rM;x;x_b vȼ?++;%ETilH@e8 QkY@Ƀ9|M{hў=â84Oޢ,=Ȓ Md lj2+-"NeRB:j6k&CN, ц;C<[8V)M%z\0uģ(r<3{ZڅhCUHz˦Ë5Tǁ"T"A>[,q](u؋`Xc&:HQ&9$KE!T^=p $/#j2fj@ΆhWi˵)IM-.Rp:=1X4AddֆC2f6{Dŷ-X b XFNQP*xYA{ "#`J:fxx A:M9MaL[b`-(D *Z vu"JpQoI\LڛsB]4W>/dKϼ[aLH 3:9,zJ\=[Ikzu uCЁP[}DFC^\ur ^rRN;މP.yc'cc9C~е$~̢+֜btu0K>ڔ|d|pӃq`gq.[5>/egBK,G:,,@sJ>K,h`SFEL1V"`7?:K^uw]~\ecm= (31 3\/oR;wW?}7nσ5W,٫B o/}}0tg*?">p6}Q[h #RÍTǁT&V⠚MN?SW1d %e)d,:AL'V-)3 &S `4rEZi"T b¿5R@M\J8.p $XUI$tQ20^ Dxjl֡.YS"ޣxoa՗Z}//ן5 G"DΧࣃGFgv]L)?X<:ʛ*% ՚1wT)vVo[ٮ+q) :$Kݘ_{խ7qЬ^SLIi֯+VBEaa;*tʯ# d ͿvsB.EKٚ"Y캌eR71l;u$%CZ}o4VR򺍴-1!w$^ >k!Ǵ~ӂ0ϵ$0, h03(RHE `ܒ@uF,[ :6>xYd&'\#?R"M`}}4B,PJ>[PJy5iNΘR"Kƕ^ ))P>(H ī I:Kk93K)D%1L 1Ou& ld:%m=y8QtB(qf{hW- 9~}{͆?LƽPmO!8$oӑ Vء##c-c]XgC0v`MT8͹NTEՅ~^]ejVٓg;~@MU }j3MionsWgpzE+/WlfuzI "qb4;a6j͙DULIPK2!b,9A]`mMv,+~buo96 qX.Ɉrw{U [l&+ղ|Zq[e r5lkVF %X}PD2V`EYP'1c-P(W0#B~}lrJf=0s#%2Wm cL<;?WI߮H`432QeTsUD䦙( ĢhM5<5^{eB@x:?Ҳ"$)ibHH$&|[@e.dWJR@m!(H$(GsW9;:Z>Bsf9\Zۘ*?V9T9 zWzemnXv|? sϊЏufi%8(Buj bq8br;z Rr*ADi?!9_{z4)%7fhC?G:J$syǿ-h]-G\AjB]~ΔcH0şRw.FU#ϗ8ѓy1Х_N]|oUkdh .sbF?./Ǻ̓<0P+?&^D|\r\#G('|Dίp n}=-k"ڎ.KbU(5nljQ= Lje4t"KծnQErOUZ56GXŸ^6 ԠHNCཅ=jӹ:OtW_.pWܺlfÿf?D{2l ex^V]W]|Y^$hEeO-꠯3.X}l%q|6} Ӄ7qfV$1 Q kp!LYJ&r46$\k/f覬S\ x.BFN-BYK ĝ[Nc'YI}mE`] v ̷3o,$ίb,BYr܁+-6]S:#&'HJKҭ:osoEzT#=Vwkz#ͷV=t `;tg &AUx~R7mkZd2q K͙0)Y]x1e0lHcPGEE(5c{5)..J\,em]عb"q)r`'<]|AqC]e!-e+z4>!ffqI4+N1O-y1o杣_(iZԖPK.C Eڟg7ξ9}SzÀN_~5.΀ǣM8ir !l܀V*x\Cg|b){LzPovKɢ@=?,>{0^}E[ճ?^ sJ\JZe =rRn qA8Z̑읕*ZꝫOz?F.6h)%5U:@Fg,x1)A +4\Sd"pC/br"H|c,AVW0n*Vk:Y4݌^RE压^ubS/et؆l׷-m4qȽڑڜɂ/ϿZU~FM?,C?)0*+r tT XU2j49=CH `̒ =)`@eb^D" Τ\&!U[2V~ɸEVce, u' O^ZqYoP\[7uq /7/ 0Nˇ ӑL-ThC.T(YI(1E}2޻[0eZb(^Ҽ7Jm!ad1Y@)xsH%v~WKըXm*KmIUYi(dhPd2$bbA$0ht_0cVwY^8ߵK㱎ǞEJ2ڝ9рȅ|!餹 C Ipc[o:N8Srڻ0ŞVDT7.lc R: f V:S2pd 2'HØʡoPy@&K9A7l㕷M]de,q*:d+aQ/DPKGz,.i0b RѺdb"Ι+2"XƵ֢9Yms=qi6AN_RoRa[)SxaoJ]tB)\A'z.J]n{wp}S:)cNL&g{: Z:{}{UvfLo $g'}lh Y8.[0g7&FZ'P2c@ -0\~!EڏZST3IBLJ.cp La8@dRNh40ʧln,ZyU%.^L}ol,ȹ3\ >t=$wNm̮wM\06fvuޭݺ/]aw-.ѿx1ʕa|c^t=8t7+Ѿ|:pw\Zp{C# wԬ+d~w:0J5zK|p:-˒!ޤ`Ч| C"i ^FAba@eHYr1,napPR)eX:0t.3[dzQ "0vcm4im=v7^^2NKdYM77'RAs}߱aai-1gTV7x__ 1fǘc5˺-|u}f⍙1 Z~ګ)tߤ^Rk?Eᓬmkȸ^A'ȝ$)$qQW.Hyw9VtPw,̄ˉ0W[{B_]9SO ѸubuOiKHز@e\ọR&328pfu,x5ԅJ+ʺL:ױkq,KzyJe( t;cJ ȁ20.J6aWRjÅr6,3`אJCB!5K_6{7eEzY %eujP{ qOV3͐&˪]_RmnҴ`lf7n)  iw_xsĽsG&'>om BF"p}W}N-d \Gduԏ.hWanP:RJ[y2EGnz?&кqRچQ+( b1xE2u`l~b"NP /T,Xn19R՛EJ{AAH˲NnDR Keɦ8fΐ5 4IEĐ {t֥/Pq#|nGJ#j@i/Z(lV9(==Y?($ _$\h( S 7!6L5>k\t(gP FAf((Ŭ*} Y2Qh#S2uH6 ]Ne'N9tZM[keQqKTJϋy7D]>$ګ;tt&tm-:% MJe( ERa2߱vZP."R:u^4mɻ,nc;~?ʼ G\}@+U>H.+Ut}!1"ѐ&W>H+\iLB]J\%Z8ma4He GcyE pMUQ:atƐ`P"<`pn#HlxLZ~GELToeF=Y!}Hrh%8ICĄc0/SED!JVw,?yJ5H @ n4:r3["f40Ɇsk 9ʌtL3 Y,fz[+\ycZD '2<*I\͙Bs `GSxXV z?bq>ۤI?&iIfX]OB ŰV?edϕ. ?LzSf n5wNPVVwS c6=zO nFr%4qBk@hJ)|{?#G<$WsGAZ^֎QX?qF|& 6&M{J84N${~GmȅhfNs9(nO@^O]oMHM687ԋ}7MèOvO%xstV~ܼi1$DR<2"\q$si<%W#vuc\+7;?xLI]mQdGsK%fW&02jn|{i8m˗{\K E\*+oO;x:Z?9ʣB4뢁#_I+m#vof<_; {;7"- BQ>}4(s$ ϒQMwz^V(luߧ_/$uHܦޮ6gyM6ԫg'p{A'Eo`>8zhV<Ƒ&O&idI'HRIY "f;f4-DSs䖠@vK-p+6̌$%V rđ֫;5r% c}vյ5bKۡ 8K`&5da^ƣ?|@@&d px)\9>%buE7Mx5K A!xZRԌL^kR]i*iH&/k;hӊզEG Z_O߽uDPukb^!-emg"̬אc49;.fE2ɢ%4:¼sDϟu@mƴzە`3,||37_Qa@gg M8ir !l܀V*x\Cg|~ɯh){LzPo)Cɢw=?,>{0^}}[ճ?^ss\JZe =rRn qA8Z̑읕*Z؝ OF(~\RD}L'(dqJԙ2Ф5:d9&u}N J\ᚘ"ېx0F#d vsgt3n\Ŧ=^v]5:y陋cV%dV_5DӶȼ!6=iGj6Oyĩ*;<0oa9dO18PYQp8mxFǪQ1f*GJc\I*"1gp&E͵tğ2 ڒtK(,e+ ueY;Y]o"IW?|YjGn4Z4۞YU``\쩖UIVdVD/2qfTۼܰ[,wSw;,N±y Fklq+Jg D4'OLD9A)@ E's#ylwQ(p'!(fJhm\sB&J-rvkl7%\vkqV-kmi"bxEF!ǟP Hdi/A Iލj]oGdMoIu2تMINPy2gZ<(`) k2Jxޣ@q(x#-U,3|d@7hdqq,yDK VH T[#gD^>'wF|\ZBЄ&a^Coǁ#g uT ë'«R?fO2bTcm9prz'zkG=zs8v|*l21A<[&eGRю<I>DByʃDz"(@ Tdc`$ɴ#ywP}/~yg]׫]0HVҦy+LXmBۻν~ӏH1^s1Nl]`ruQ=NpEEnlb[}v>γ@KtZnhb1Vij|Gƛr㙅Kޒzϛ8!P]k}wsorlJqÄt*>P'Bl2,x}b.KBakͬ>NaZ(zg/'`:'.:k rpAb"&ygLD'-TȬӮ;qyˋ*9Ky ٝw,-+".:2: i~  ri6IZPR<`,X LݑGwqG21E"!׺ >j <iKRTHBk} LPD+ #.e 빱1$ (ԡ'$N-9:yd}S8,*,^]%:'ArSXXTp@qYsZ𴃟Oc 3"JCޒ@R\V3b00kbLriO6I] I (DqDE .(I!FIaҡ* Z6l=j" nNx4%{F`F#̔Ѡro:{B4!~ڛ/:Zt2G& H 2ZIީx4â0! ay: S4Hު< к18(|k6Z::G?F2lNЄce5r6y A:I9IaL@60&JjI ȄbZ2*3$v&DT`ġx&:7d>+OqH^".VdX 6J3bUa;;'CЁPԯBy˞T~`F2*e:aRDE$\*C~v`m6n6e3=""?|MFi@Ņ/8eaW]1)) k6{|BIþf}9T.X]ۗ?\-sCHП>9mҼ`6"F$Zї@%\Vf Ky Xo5\j9^ K4 T" Y4]-K@G/})` 8e4/&!e/[϶ UOu狫&[/v>%|lg6w)'R # ",QY_ZLwI!cS @unZ**#-3\8. 7eu1vw0Ŷ3V+3L!;|<(Լ3F <+%p֡IObb-9LZ^/,8ZJDd!Rsl{d]a+цZ{7i)rxp[Mq[rhws"A@F4cgB!4j|!]\YRaN-u(%(^ k x5CV0U";|+?ךIB#A̓&0l&MG919{[_6yW-^\炙t"Hsqщ#g.ec8[|g3NCfU{tT!_/!ivX>G ]Y.OU Myz3^6㥿ƛESr|\~{z;/#wO_D-Nj%YllZպO<N)?f#OBy2Ļv:xXP"Z!LrYH2ڜN~K2|"oGf\oFfZɎKeTz\a)EE8\K(Bκ̶ʣp1"em<6-uUSmu\MbZ>O/v@q\WK?Η;iUKd8p>?-'Jau?{ȍdAx2C76qO4.oDm!%gIBa uO),4QÕiɼ~P FyXKcSٓKh2כ揹>>举OM̎sWzgiϳ UBW5gA5TTǟvc[eF]^frwn*0KNnRnĥTL$А@3CMP3ę8V* Y@gmM^;Yesa麗=Bsw Eha5) la<û"N&vry>rYN27Hf'`N0,NBĂ + c(yyuaj`}W6"YKj۰v8?-m,u]b:vj-NmQ`zl$e4}Mo&nSNLDX;6Nwx#3֪Ia\8O0Aup oc՘Vleyy؇-/.;B::ס諞ח}t(o EM oK'-1C,0s+M1)FRQfB)8bK [hZqv;~a<_۔Uxc%ٔk[Jc72UCVK^Wum5K! >u_)5W{0R:X~{XYEpǃ5gk^7?Wמ便;EW{PfgK!."zhzq6-LudΆh8rĬ0- Rx0R!ݨ}N<53$ml%tGsfYGfDXpƙh褽h`:Jg@rrIsp-Z.W(԰[z=5jG/_mT_\asrJ4'$SH7(EK,(%8Dp04dKa::rS =Ή KRhZ@ H+|wT7M692%'tFjhJ&*F#r(D;MM3 R\C0*QYíN*d#8ṌG77𡉗RЊNPJDDѶvDT`D`#!Id-nW|%!z11&)r3B$487yФGDhу'Q䝵Ddvp4Ef^H14  ͍cP±KT"ڔ`8r YFKP}%v>F܆J$eAIgDǔCkB2z ;f &|:d2/84׭SC.w%Bjx9V>QʃQJl$fTH1)!(9q)|$qI\#ŸvT. v "Ys-lʶc2}>Ω)c]#L}-w-a\Wnx;"Up\@ /^HM,LY,^4Ů[EkP s126@"KN\Qs(J[2JB ")cq]$ElKc+ y :ΒwTB\1`!MM(8 :}Kc[Sn5Kd[D/@B]^j0g JC+$d B'vH -ϷE1M ̈́aK$-0&lJ L+4P2@m+XEer  .9f %a7i,X \-ERP6Z\*Z!ȮLD)ԥW㹛.K\Q$"/=ZQ $WEI8PR46Ja<| $:E2a+֢ƛ ~\uXgLӄi ~l nZ1/Yfz̪(Nc@JA; B u@esa;sS޲a䄨G[-kyר8MuY D^x1Hu@KyY=Pzr+*Z@L.H$UzD>L8&8bgP촱ڂ|ƂBG1 P$bDM W$D胕EJP*w6POyq `|Q@VǃV&jxK[`G[.clS$ 8UQ+ɉ᫈;͊yn[&sJ $'a O6Vx}lMyH! .p /룫x* Ge&0!oZ%tMކFL1Žn: L>\;e"nu@xchdr[< N @I:e#s4pqF}Nj#:MLĨ4 Jk$=|P@zjUȰllFVO:+lE]("Vu\OMd1DN!FolgV0 .T4B!ѢaT( Ed'UA`g rjKc-и ` "eYq2csJ6HiF#i*\#|r,%&}*` WMIB1X>ݲFr}P*忒 4H*R3QUqZi#e7 tiYlÈP/jC ƅ8~)Q#2`pϛrۍ9ml!&IeV5v:O&".ńE@ZĔpIDa:ubE S@$Ti@u!F芄ީEF8Ø`j?®z4.T8/o-Έxy e6A)-2ኲs3Vs颇`jTrpQqhND;a9ޥz GYޥx}/j~~\JTt,-Q[Mzb38RAb E+ls^k[yW,\uM*{hR,bWXAs)Z%ͯ3n 7(l&ɼռ :O=`>wEg_}gڛQY>,*bT2%ZXcfV;8[͐xz'q8-Bקޝ 0}=4Z$>^:w.VjX]` v.VjX]` v.VjX]` v.VjX]` v.VjX]` v.VjXߪ*b[Ow׿#\1U`8?SD=@t`InL[*҆=Fhn|mkljshʦra#͂S+FEJxgL noHJ].b:fٿ\9Iiiyz+q`gF]DŽͧwc&9o'9.rqk<ʨTJHUa!~My}ykDu]+}xރO=4?(5=6VK=lixg854mk{e:]?(DsoG'6央V;]k'l@C904qZwu`][9^puA9:]tsuA9:]tsuA9:]tsuA9:]tsuA9:]tsuA9:]|:"9?u@N?u@֚'Vz} >}$eY޼_C+Ń Y 5/o X4UBit|PrFa37lY9!]s5D_^c#C*;cy*&j!bVwo-c?TT=އU{oo^"]$-_  7 sVՎܬB Ha,t\J,*%U#:/eV{k*uκ‰GSw4_NH u?D+5JhI#v&yܜuhmp/^+rF~-?\ϧR~}b<pKѦƵ'>=3UǢbof5Y`O,8 8 xqE@4ېjMn[L0o?3F)_%6csi Mυ5n,g[% ɹQm׻/HˋtNHb;2Jh$a&j9|ą.I gWhC[HdUh! &cRPڔ&Zﯲ> ǜBy%Si;Y8c73gƎ|\>;ڶqֶ=kwC2:hC$?"Yv$b0W_m156͇U4W`@\ьVY,rM.,QpD!isOΏ·ٟGR#KdVqlFt3gɪO =i=Ҟ|= A~Zɚcc,ԍ'=tn'eWøz_P+E]P JwA.(t]P JwA.(t]P JwA.(t]P JwA.(t]P JwA.(t]P JwA.( J{/hzVr1(8TMZëyB bkb~@t ud2[\Wl2;)o >,9X_<#=IPڻgq,RYF1+m#IK,ܳa1Ӎy6yJlS$Gl7xyLIE+ 2YtT 07*<',nPvf%u*v[-S-m* {S>ÞS[=6Vw~;k.lm~~>7#^s#=r\N=\d{҃߯ ^nkqA&{l0˼I=o;Cs8|mm…$3J(ǐ,Ü&.BDB͘|ZaN)ۢ~yRg#A><"c63kX+8.\Lq={^+: Nj'GCik=J|86N-±U8 V*[cpl±U8 V*[cpl±U8 V*[cpl±U8 V*[cpl±U8 V*[cpl±U8 V*[c{+pl_'s;^~bJJj< 0R̲{m `r(~J"999Hd EX*PIMy"WRMnzM55v0ѢH1KorU K4QtLlp [䙰ELDZETELEL'Ex& L<{ 9,eȓ!Lz7E(H]Alb.pEE|v^^cb[ _^?ySKT׫ S_ S⭖W70om`ǚ!5lCLfc4趑YhJNѤT:(2I'e*'ˈEeD\AY)FΞ_ UR#)i+ M:QBH Dsr+-1;iIK;$3ٷ[[T,W-c4/PeK2F"ނ~ŚBPQB6Evv'B=,aWJuHnR}J7{ Y{q1`P R8PZ+j}w9We>6sXJ&`cl1>_sGϻ}}ŬfzYb=!ᧃҒkO}+HT} .0'`OF]!Vjlzl]yӏ cg4b)8T#TX5֮!G!ՈPBfpyԂe3>`H'xLh4KI kbֈ Q,%E[X/ڪ^:\y32GqX/inIv =t?t'@?)q*WK-%ݜ?o3?6Ri<87&jh-&款y9osmv8;6j^3e؜ o)!!ֳ<9n>( ",yE# ftJHƽ[Vԕ& ƔIK2F"^ 3xeR11i5>jKs$vgnߒ;8iT)ֶ#l2ۮW6&\}L|ò<r^. **lb)0BeA'R`rY$J㮦`Wx>hv9iꢩ3\JLkAhô؞ 'Bk4$Ԇ7Iq 20n[}֖f4Ta} }L48cpIiapyDad9H4DT})͌Ti՜kr[`b5&\O 6t^S\g>RX941$d3V\u+ JivFH1G;EE&F)= Lʽ)s(3f93/ ױBC һ$ r3_1u>BuUbe؛/hV O?~o0<6R=/Lujv~a}L_2|8q5ʬ}#QQa! TW.K1N<`L1PK]Pc|frv$bWw8mvg k3 7\3IXA͎sȀ>eoeƟ[it[:ْugZ"h9>i}sׇ1 z^hFfO7_zsoB<ηtl|f,F|ǁY{cs֜ ִ5ģmN>> _ʼnݡ)ȭ!I] ]P_蝹XAz uPAn%2&P 6H?R/("lP$D9 JTdʜ^9tipZMfig|4XwȳCښ6J3J\Ӄ FL!LDO,hn0Kag4uYJ/SRQ1.'*-b'' Ë:_>~gu6!AÃNs{}e+nkKS&])Ӿ*ix-/ʁ·L(:^èыoz͛kыw,NYemz 5w&Wr"qqPݚ 2|#CEx\\I$f-`n1ѓi?~/b5cz 1&b:ȼNQ >Pk8''RbY&r=VLQ'AZe2 |FqDÚKAą1PRx9 A^( %|BhҜe.&h%Rz +'0! aY: S4KVTV9kˑJDI,!b@,VLxaJ;E #\Lu:i 8>b+O,@Z'Z%k1RYgD4SR\]*L{1p:q+ Ac PqA a` &1Jq+c$cs[Co#ifPx@G; ]Ї)#eFZ Kιh2RMAΡDc2Jp䵦V@hŭh~oJPq{V9Y˴&c6DL,IE)1AbX'-ఆHӮ== (<-.uchJ[L.W?f|`PJ}gQ,ce0 -8Å3Du ~|Ԗ)X `>*X+ ?ǢMLB3ccI1&q"O`G'꜂H,*F:腆~  _oH[WEWNm r35^:y&S0Abr}-PZ/=Si ؙ0yɋCZr-}^|m8eBd MaJ@ wZI-r-_hl x!(%2@, XCħyA)1{uP4b( 2Ա %gsxH<&F ֆq*FRiEs@MjY<}}#2wQsMaWEЬ=U3̇)38y0 ͨr"PBT)+#0$H9VDƎl!"9ѐKldhښ"qI.I(^2 BNB(Ev ^g]y -u7ϮVͮM@zc);F(<\ Z#!&%1FqrQ[$\ce[UbH R[`g|`lКJL"fhUl̔M=u3cnI!A+9ݸqkD@~=+!—r [/;:wd$_(\;|/ϋiUkdK8CJObIoQ E{'K.礨N_{7xx qiW/h1t.8 K^pPvϗq#-e7.f2اuP!Nf6 ?u_sͬ]MO_?T ̂.k妕el{+ ߹[_p*1acze,Zb{շZ NS\SeB2[l|YF,XPX(6"ucxaZMW M6v2Nos" Y:.tu.=jKrK[୦ۯ /**S4.g@\!AY-;e5w`4 x![M(˷9)io@@'&H꠵p!&*UͿ ?Z[ PjadªGլk:qUh2ZMv>1qgS&q@B٬jV8Q _~yOLr%Ay;߄Z#\@H8}񷷝[oK(-pmc]clͱ8pƯ۹"|>[q4,~ Y/NA*[mTmC|Lׅw'ෘc!MjYQW궿ml}POe䭺vx'RZX" q rsTh㙞5 ?E452H y\>ɇmmZw*PoY+j*-7jx[%3/y+oB*~gTdyV;[mၛ lrL8M*cA+bo 2~S2\e&s]ngz)Y&o,eq_@,@ɂ#QhON;x]EpbĀi\q<(|wzSjt&+d~+AK2p] ٢_R*a/![oWaJ8&]ؑɓՎPkZhGfx7oΔѥ}->l-*!ƐcJ\BtQHB 4'ǒjM%d]-!/skh7$1+URQy=@.Z6BZ.oBv#nj[ي/,f3!]U2w>,dT>vsIy{1{6uFɴiPu{WC;Ρ3"5 \JMgr՜Wejt5m{TW ήH]]OO&WkQWcJE̟Hå'XcLns;K09ڍcoSWQ/6A]V]DO^ q5*赨L$Ĵ=+bo!t}O2SyA?{9 .o'3EYx@$,}NA\> I[ \i>xJ%J)8Ny(c5hW?c4pq8x=:bcSȃ^R3@N7JqѸWam99B(Ek18N :&GBDX &g F8%kmua2z4LB.rk9%"6ZrC c;cL2#Z+'#7Z)‹BF!"8 7'>rr(LkN)/ y["F 8};`RCk՟uҺ|{''U\Zv]ֺUOU!|Z0jra`N*2 C,iM8LVgzɑ_.i&v%.C,m).H$4X-_m9{fbCӡkJ!J;?p<=ۮ|bl:Nc2.77R)qsްhL"EFmrWRfWLv$tZ8VVyO(IZQfU: ;qp@AɁgcKPDmV[HV1wAǬK̺+}0IKBXz$&TRfy\bһ^`kQ gKFiQjX]/B ôF-L|?T󯽿N.M?LO_Ԭx[=`ڻCݤ{|{9<4ɵp>/==֥chwaBߵy6NR-P kgNg _Ƀ.Y+޻>Hz#OǛ2ѽ?F{M̃%((%/.[0+q2v5 qߗ/k;[>Lq<ɡ릝|W7C2uHJ +Gq=FJ.m<_Ƥn/ٿΰɎ!pKRmddˬ]n.i#))^DNru2~ȷ_PoUE;}od|Jm{Ԕ gєp`;"IPC&Mڀ8dG|R73k֭bv͉;> |Y0eZ;|YB+B@K~)qBӫ6F Yw4&2%g>0s\|oS)tԏE)1iR7\%F摙hCJX2n!‹()CUwh'^Ϋɥ=>ov;gqejvS-_uix3me7eO%];}YCC>LFos81|̎ڟ~ؾ43ݶ?2xSgeFn-ͮnOzy}h~/%NcbvWT~Yy-ycsҜ=liBy,y(m~8}8qօOJGQVKQ^[cJ5'&霹VYSY՝P7U)-qiוͽЬR>^)@Ӱ7qE;>N(et[v)w/7ɻfx4PT־hx#Ž^X:_-_gِxf;Nj^'_eʀs^0'#qmi{h;58/hht W̕>lkk( ?s/:jǍ3J(IIE< 1V쓌.' Wt1(!$i|9hdhlc)C8J+x BP#rft\%ash&rrz;ɸP̧-$)YaM'cb(PEtҀYk/q٢ 3*|rPMS·S %Ldb w>!XZX"˄oAZc9LLkZAbA "8Z<(2}BB愳 "Ϥ)+!*s>0 Mfk`6 >Yo_NS~KA]h{ Cl[R\MH7(;EXvM O۔0a^ 艜W(dg8' I`!`" >'鳗Phd'^s+'ip ϸIcՌk lKC?v{w-| 3!m.ک;D֑pT|}Gm cjE6C"}S; r>kݜ3vɫS ;E}4"W-aTy_co0Fmj8vdqA^W2KC+f3MsGݏg ]r-CǾ~<10 㛎oΒoֵ5q2}rséGd=]6֢m^Qi5>nxcS6:cg̎1;<cxn3V[>1O6ګ)tע^Rk{./! OO8]9Ix'kϽ:xG&q|yM6fDTx}%Lˣ+?`kBYtTZ݇pa^~d4c%+e4!7j(qh7lx55uYD 뤳;7ӡu]SP>5NǼ'b/_hǿT I<(XaR9``RKrVum#GEȧ[tBY| 0.fr_MdYv2ܟr$-VKj NEvUz9939&L|2;0U=U$Vf=Udv[c׳y%ٖ¯g4wzza)ǡCBPPG*UߢN:jz%ͬBJ:*br _UuFzŷNp]Ci[n¿' zHLH| ՕоJNr A;ӫ:bSyPNVA2BkA`fNCO"JT^ Keh1Ԕ?zeZuۂGC#3vxϙEʾ5\R^P j(˾QW`}9]* V]"j辫BԃzL3U!Ũ+"suUPpPWoP] V^"r5ח WWJ zJU`p[}w'P:Zw6\5w?~Zt_Ï(>21OFFF7wvsBV1/|=-@~•|_ƓBٜGF#ܴ+/dḜу}t3M/ku)8Y -B%U z)shY]=ca??Z{6J i?=430?g3yݽT/Q_Tu'\e_٨ *)XFYca 5M?<>@'^'klOj)cNҠU2W20[y ^'r^ʷ\!WK ݒ#*W;dܥ[r磇gVW#ܯEDyu<*엺g+K=`.H]-uE*.E]j*TTWGc{UOڛI'X) Z,׫}lj=N!x[1<3P-ݺ0kr%Jȕ *VكRy!ppo]*ɜ#q!&c9E NTMBRFƵ.EpV>eiKIaÌ7ܦY[#q,F3'D] g+"Wu\3rlͯ ͸JX6(\/ZK{Lmɢ N'eǜ Ms9fv7ӏ[U[I[*U3WMȠ!βK]-Fwg3ν]v>.ZZ-g^non̷O{źB[/-S"\^ fF6 S1쌜ݸ3Lԥ"f2+ GHTȗ]qJƁV!,x?}->raemP6R$@+ FIu(ƌDR&r%SSYaPkR, I m2b!Wb6js,ʎe3r| IVXwd'2D DO:h5C\&t mCV(K5TH!Q8ysVwۯ#sKiamvO3Ysφ/kO.1}Q:@sA*{Jy#>\RçPkTk 5|b !^D-keJמ+ @t%dR}7$w=MVL̢ĐGrBܢ1(BTԊ~j(S"!a],#zYre&DX[cLས@$1❑ijMCo:D}Kc]#ԁ1~u<>An"hu1XgN%õ ir^~۬VO)bJYJrZWӱs2~s=l@*CyPF*%4Yc-,0X$xI,S\ B'!9#7sَggr]f7_ v!狭h~q*Oz+bJSOouWvw([tI[14q 钜*NB芜$0 mf.Vq[emEvb,dYuЇ'Iz0*_UӁuFzwp]CDѶo6ddr{an퀴΃$Cfy?+7}ILSa# A;ӫ:bw_?O=/0 k$x0J[#Q2a|^,؈Nb#6+N.i«8sG-ukOG.}=P dPډQ]~,MpB&/ ha/??H=h?,g2-uwqۃg nh֞O64 ̊.f^ 2 b:ey,Ґ=kPFR虣L1Nr ) B<5i7WegݚEMg*էo&\ OH `O07,AB[Ùk oX&漼~0WN9Jw#k< eAdGm:t3tKAImM6[6GǑTTR}_ߥ }1B;Z7+-5Ih}v6w̹>x+]2B>fI qe jFfPј63Xz&OQljrqp>ųy[rX"@'z)jvnyՇ>ʅzxpjs,ʎ!xg=J|EB8Tt{mKoQJDz6BÔԓgJKpS6A 8Ft*ȍ+ubI8ɄY2}⮆dI+NTa.}62}l@ZR 6O|7\:ќkR$%Hj +s"ƽJӴH, >/sAG_J9cGot$MKb ˰Wm `Cga~n?Ѵ %`:*r-F3Ȇ4^SIz 9ұ )IifX,fzZi _>> t˜Kkv3Egj{ȑ_ib sw ;{AEȒG-q߯j)Ǧl9 n])Xon\ʖDьGѨqI!D>P+]l~IDo˚ߖ28Rw:OuhF{c&Ϸ,%IN|vm1 MW|-8p%Mh9B[I(w( lXnВ4|>FW> QZ){Z(K\,wĭ12"\A(-iljm2|Tt⁁+5=@>f K][+Wplt4Ne'*7hlμ!o<n5861<8?Θ8o%K}CAY.$r,fJ *%g "ȪѦ)c.8S  XLϕJ\Co39^ LeXM{.Xa+>"x={[ RHĘWù.eF'#%5,ĝs6םIDnI ϚO@D^H%Jq4kKg4+yt`V˛cY'/}d>g}, ݙH']>yKM].U{** }_Nl%D~I qZ+0c"q6#@, IE !YMh@ HFjGz\VӌbfƒgP\徛_fesqv޴/û v#L$]Jaɯrc^ddUP]BC!@ݵ dV L',KIqFQhGɷcgJ^c,ZWtk.池v5x(jʨmN'BdK)AY28$IE'lx$9TìE+rUd k0hD$Tr05HNuwxXMx9˅G㡈*#q@]oFt,YJ˜IX֚$ FV%e⌈" X/eL9y ): LsE>t|o4z uJʰMÓ1<*ڠMN^1T֕BŋړWfmUfP:PgTsft . EoͰ#M o,KKwo~#׫L]Qgn')LJ_>Yzk{߿_%`.d$]hIiYwۑ7 2懞ЁЁ2F%( s,ozmob,:y-?,Gp:e䥁Æ<š1ni.eyZ-"];٤])`+t\--הb;wf&VO'oG%Q%lԏVVqB_?̸$%n1VPDn+oȹQBfn@P,1טꁁ k1Gb"% $i9ʄPdi09-cBUYC?-ZېBJ QzHȲց}VdQ(X{\"Kd28[}LSQLoNu:TǗ1!Z/|?GV.  ^-Ӟ9@TJbRFN4^Z a-H@?9#|ZneL0M&Oʩ2Je , Њ_$UB %]TtUʒ JJ̅7TI"w"s8 ml;!4r\?B|6}Vq*z."vEVm?%=>Oy#d}36hL":BZqC>Tpu7!xsp|YnTZ!t4ȨJڛj߂7cͦyi:bkwmbS,vn+L뙘7J1hX%5a#VDs( M g0eyZjA+&VeIecJgZiT҅ҳlb#FE`vz#|%ͼֶ=i^\Yco4m;P"D;z&v$smw$t/fcWݱHkow,Rz=lw|\~8/}9q vgBn$)#Q{v5/9#^OFq\L<`z'-&隄KݰB]%gMԟ~4k/ `uZF5y Si"4E!04zزZc1g"(|Isvȇ'Ȁ~}PkD21|mY*odB 9 &{LdA" {Gif1>[ s /vw \~8v, 76})3>|z=)[.cN%R=h+xY !9(\ }ωiKނ,I_ PtW`x`pN)L*88AeG&ZDIupVFQqbʎªDDeSfȱ+'~ao_ŀF__N44~Hc:f (܃x bҚ,sR24B0r)Dd0^ ﬑~VCbyatD${a 7<]`{H}jr>o,l@fs;{թ٤smKؙ7{OaD.r #5y9xmwmH47m,̽ w{;ؽva'Y<<}ݒ,9-[N:EfŧUOY##D2!bB By#1I!b;״_޾fGν6(s|x8{`ܧ+q\Bda f0+m9H/zS 3gD VHԞE66C5*1C(lf֗r 1ݹU<yڙGpb:VnqXua2% hh!9d{Z%A&d]Dy6`8-|F8ԧBB_ą`' ;tT1 ENh (; gT6=!T$:9 5-)MR{[QxwYO/QBt_zioy-=ZtV]bH8͂iD qM0"GWc d੅uyܕGK1]|M:nf|޲d Y-A14d$4$eRsC3ԆǤA`m>w>6E{Ʃ?ͨ{9csD_X@ZG`-1cW(PmCU,犯~xVQjF /ƅ쎄>V˵nIK`UXt3oPdi'C sTnt̯78jJ @V߷r@x֔RoEIC(jNeǿp[0]u3.N}HjaKJ9Mi4λpY|_<=~n|g`ƨl]5OD 8ta[x# x0zYQq$spQghfʏnfsZ֋cQ$.0Kj<8\-Pp-=ae!J=-{V/i~HkE]o]dGx>YߣB\~a֤>C#d_8\)/9]=:[WD8J{Eʪ:26MGzE^$xswLyޠޓ֡2U m|>g6n?*x+o6 oD0L`6y%\ o!Fv8a.%e1E.%)ZB'@6TRUBYbl 28඲n̈ =br4 w6wl>HdH."BOq9>ttI_pЭr!/oߦ껈/JBFj՚3aS*ܭ}ŐcҖ=xek ,I0u-B.HA3V4Q7; $8nB:^F]HWE9K5]goV~0> ff9F@2+N1O-yON&}!U~w8*7yČZb7J- ȪE1%d &sU,L3chNҢ1[L%\DpM3 ېxp?$F@ FK0 ]sj8'j6U< iqb|\ߞʾVgr3lm~%pY}CLЃ]$jzjgKb~zT+|yh~_@s(1*-8"\8Zt𘍎jF&ǘg-3pA' "1gLkp4?ˤ֌ gft ;CuX^p6)x۳=ظwGngaQj6h40_V<8J%+ LGP 94Jh$ #'vP=A -M&ێBC,Xcw6Oq<ݙtMZZGJ<[RjLvH?1]ĠU6x,WMtHgߩ>JGHqI2!C.K2$C&&HC1d|*@Fu2ag٭~YhĮPkD5bx-TRȼ$ )Re*Fg$CZo%طYΠ+66H45704 ‡#;,jOzq>%_ggRr^E^/>J z  ,cH[&ut3JqW^<^<}ؙt̀.ᪿևGPa; :Yob6z#~|GɹE_*!Fw՛Ma2 `t1/'i~%-M}=}G7MGQ']MO~Bep3MO.g|?|O:h̛R4y1kFVRHW7YŃ~MdߛA_E }CHH۶ 3SW% [VF7,W5oj1U`qF./|QˣUtUw Y9,*>U1.Ty=Ԯ@ o8ͳ ?[}t}e۫oOx1ZV7³[:lyP`238xBt 7wc lݕHYH[-fS}^ xp2/*MeWez uwIw|thIL ^} ]gg4 ` /뇗uۺ44?[ٮwƻ;_y0Loytw!{,lLE;:.]j֙Xg#+\DKtny ߫0~EՄrm e6oj)5${{޼j;فy$:NFAcFt+ [˵3ŘVa %h?F+BZY ںwr BJ;MAZ:;IVsy`G&>om # CF$Gvv۹E쾀ώS'hnT=t 1-u4z> nD%Mk9]d `% w^,o)K?W)7ܥrLx02$X0' JJ*e>H/Xu2,*{YVʛl͒,R$"b{C?h w] ް­3=¡ ~b5W~Z. yx Ai%+[j|4Xh=ux}_O\/jH R܅_8̵uL>.{>*7/[;Ŭ*\G,r @Ds.3&DnK<!#>2jhhleZ3Y(z; 0Eg⇫Bű7^ȘCv&ҊVew~ ZrO`gQ+݃2b 4NFc+ 5=q1W EK6GJNüB#oJCLFNqCZ)TY }L.C4Й ~$hv1.l8{R$L tYT ZEv c9!%llˊ,6}-K;KHMrJ (SFpBRNZDOSL8yzpPdy!ƽb (,>koe Ar6kN#ٿ]`&Y40 .v `$}#INjɒd|8]ux,IU(1A`*0p#3(M 94li "wLAPu)a@@߹)ӌ|9@4ףY/ߣ]`MM>S,7.l>eLPrId61`F`Oha qJs?=3 Ϗ<GRD6:idOc 6COP j$sSLD |ux5ݏYirO<Խ92ӗ<., j#І G?瀄|  gW|ѯWkٛƆۦvK>c{C+i֣@VW]^xXy(]_s &BK(ޓmz ߺ+1c,:d.F5Mh D-\w;H]MA#ҿ;TN!ZɃx<ːG|VzyD1#nJ@P!4,1>C~B0iIENPhĴZw~B3AئYzrRoau_bG rh˰liY5y Fk5j]) h)!b d o t5[8~ٴr3s$ICݦol I5`U=ŒAϣ_P,doUV+۫i'8ZUNE;|5_kl5X0yr\v~buk5yVJ]rɕ.ɧ.?շfCv-D \֏G0жU#')W?H~)N&?427+- n)+n6CY 㧋ٜ4vMwl,yIؑ GR"qբ+vA:SD[e#oe{&{uӊY nM XL:Cd\S1R(zq~S&xc˫ xL)2REeH4!Q,j\kKm8Gb}M\]ov(f}6f/FnuhFLsY<,FBN"oMl4+Zx qp)IXAp5, QВ4qЙgAP ";bJH KDZV] e6yA 4&L6UC582a}For>| ^t': ; tIOR.~|P# + 2mI |9~JN;翯x~IW्,;ddI1V#hchXƣ!hM,Z%Y0qz^GW^&yKwo9PoE#Hڥ`B&C-3s ב+Xؼv)xM;Փ~m` Qc o f{1WuXZ _]}杊*f(fC1ã_)[ªO=8UHXש'dh7P6ϒUD["Y?qk8 X514XAO6ZU(Uڜ7 5?RZw#c=R YƱX+cB+ca=4*D,vE·? ?oga*^R(HqUR%mE4FBq !dBuVBJQ0HDM&jD#L. Yԝt)hleĮ֝'b jWcQTFm3`wZ<"[Rj'{K0ڨU6xr"W caVxEKrYB25X" Tn10+dd^7a6_M`<De"v@wY Dnвbf) :bEJ)TBN@dH?TL$rV~ )r!ZIdcHZHS\(BZwHIlY:Yɱ*pqŝ*`Hb*Lnc(ϼܹ(6HS`*ō5)p\u抱xjxxSf!:юi2wiHhH)YYf6Li /u*uR?ِ9ɲ$^!#sMP5Y2A;dC~ˮ6Ǘiav6|;=Nڱ3ѿz Clr|ͭ_mos\Rc 듖F9"^ZPPOJL9s%3q9k c/}`1U4zN'0C,Јd4XƄE.Fo}}ݗc~߫_L>N?Oz]6^{]S8R|Ǿ,<*qwl|tI't// vO7fxj"+^ Sٳ{l>T-$w.w_*p8OnL 8/ SAX6 rV wNT娾"}Ĥu\Bt4ʀK}BE@%Yݰsn8lׯ,ef=MwMכ}ήԊ55_?u\8˚fqvŊQ}_Jlכxjwhy0 `ogѧEu%ӅkwC ye_zsYsr.N ?L ;yKWK[ܧD߸ A:r$nsu ҽnCAbnỹt}vA= ݞE2c[.w@Ѩ )?rQIRd aސY?›c)"] 1zϻ5ꏿ?t)h$q`L9ρ~)tRBpD.c)ryPVүHe, kшR`K$3UB%]Ge2p0 hH3N /$nK)ofUպs`4Pm&2I>Ld6u$޵cٿ"a&좻1h`l$h#KZl%ے%KF:$/Ͻ|+(DsTʃ ʢυJW(I!FcH,ȔCoUI{顅l]MU3 F \)Mg$}tFML&y[ˬq-h#xd">@P*NG7,JIJˣ2I"0rϰ<&G&RHBc x1y'|T!4/@: Ζa))ihQQKbD%BZkOx@B>HJ )WBOA/Y(^(N.IJD)yRs0&%b Mb GJm;PP_UQC&1(E Rs%2=3PՖDhdGDI;F_;p ]\r8pC0 )4j(Ť䔭Gճ=Ltq6{*OYػз}_^,S1g)PyF[xwH,Bp Rf%znq̖0T =7ii:k1z"\0@ ji.j_EL QAȬ3;-vbwpFa'K;6k/git3 h i{7W,9 >b%~'5Q1Ônϫ3`7S<+5V~3АH9w !<Bcd?nboU0ŏO֨;%TOpAd t4uK4O(֮w_6I-\Fi=<m#`۳h*$7~Z/T]}?t%lzqNhzESǞ .yZo'n(hJ@W6=VVEt 㭡 WжUF{,gp7IB~8Ƿ?0c{$mό4?%'OA"Bp-=UF+e* ҕ\m]!oPt hn:]ewtutMt A؋2ZSWHWY  WUF bHWQ!jπn ]!\IZjhj:]e顎N8G @1 Rkw n_n`B&pykW?6`F;~hZn0c#Zc7E1+]Ɏ^inh ]eZ2ZM+D Dutut8X;k ]eFtQ UY M~y*U-t~#+D;uut%r~"BT@k*U]e@N(] ]I-+ 2\u%j(]esO DWXKpjG4(Y "]iI@)Ԯ%%=*=^jt(;cq%p%SK @\iuR*KSpG%`!hA Vbܟ;ݧ Vm.dGfd4X>vӶ^P՟gȣۛw3qQ?ߜd6}kr"}~XUl¯Gqz -% Q{{'ЦN(1zˌ+_ hpXu}3EZ_-2zﱥߗՕ~~?l$`Yh*WHGZ\>Q!&ӿ]`fqk{-𨓩K?&.\'^:f=ncby6kW[M-s~1ևԞ`zvL@i*D TC10\ Y.&X_#?-{?>G["ۋhWIuc1*3b %CXΡ(DB(O,p )VJ{d8.rm<*a˧&&A/^Ceu,':iˇV d.~ nw׭:_x>2 xԟ|ZZK =zz&Q9Qhs]E!r 6'[U<__z> oh!˙ϝ-rlhC~Xoa_69E{X c8x£fRoo[7X/:x4_$ݜ|XMzr*GXݡPLzT֌NvGmѳ*a\z2 vN9LS|,U}7<3{|5fza _\%FrY'^6+ 2b3̛<$ܟv0ȡFOXU]Vtf>[f.aee-?ft5KsKЂH_?@I3znqO3ϒeRk7’i=,aUW >gl!`bR>,E+c)jo ZI4X7H*qtҭ*j1dP9_I7 wUt '4I`_%pk 8zʨ$+#>\pDKqP'eh8 TzDP:eBQVXOve'}Mrx!jY~7Ր}LY}kƷt22Um+"ӫGҫ(W˳fPfQ[VTƯ׬i}LԻOz7jHe_k C[Lo/_SA-:Qzhc^ &z.}R+e <aF ODH<*)Ջ[U8BsU Ȫ+ڶZz UC2˻(zwWq؛b<:r6^oiǣoOݰ5yNgly%˟(هď^T#XYf1i噯>~~r BA..M&gF=gjnzGzHj%-B.^^̖M>)B ,\PP-[jz5y\4;YyA* I{mIzn%KI/bA*hX#X<1Sc x$D\ 0 X 6HbjW0|QǚPtqJq!`A⥌ v/G)-щȐ%Jc cGq)c;kVnu#5!| M~3[P"*;{a7Rg-7RO DB-ME-; `jkKb iCi'%(l92\3"sLy[ΑA#QpS-IҲM?b>cJaP 5ף5օ1ۄ1xrt00m IW"ܸae$@Şؼbh$I}q-PBtǯu!ZQC7GO/~^2oEl><50ڬ)6m4WɷM))w)ƶ}ےyRA܌Dv+?zY랏ieZ6`yn^Y,h5wNÆ=s>zwT'@,R.Koy@Q{Uod4&f Mjvg -);uck^o߅1 c,c !'$r6388IFXPG!S!& A9-.u{IuK?^L= x^,B؟|W҅KmlR$Ii K?g 1er NZ ;ЫzECniv7~!UVӍGWv`uU/iaqpe*“Ǫ3.Q*x;6^džXўM w7(Ү"?EJ]B4- I4kE+BO(5RJ(gMc]Ma. ѷp"um Wհjn>E̚/c>V|jhشqz.iҨPhک@J JX7yѰL&0lG^B[4o]bӼYMń(Nˁ9E?&~N΢VxDWBZͼ]С`-zBU=wy;QZBVwrW|u6l69N G*NF3oLZIFQ55Ii\ez@#eO 's3^A b"(TJ L"+Ȑ jJ !%"Fg =rRҜ{UY29rY xa*jXQn'd U\)Dcdkr1%O9oڣEJ֪[ &.M8*MR$A18ܐd&@?k^~ɉCRQ0 $gΏZ6ğ}&CT2ɹgn gf(YY}=QY8XqʽNڙڬRsqq#X@Hrq?鐼tFBDAWFrhK J""S'޹(P= Ҍ,J(oTa#d,TI79`݈ZGl?5TPvPCx!D2i#KvHdE Fgɓu yg#Ͼ*fm-\L̐+L@X F$Aq$_8,Sn>]Mx:˓OkDZ;DqRMD},%aL$OHkM*1:+d` 4z["g̀+6(#( *DA& f~HI9B9ah4-K+I(i>J.%)uߙ8NIkGxPe-.D`7$Z^nЛqvRhlCYG E/0$S*IoѪ.8Wn4㒉1#%M/3Zn/6ϖa)Ž.0K堌l4f%n-T鷂i8kOבdFEǮ]*{UjӢZ(կZ5e.1?r I%2TԯT۴ݔohy՚ oݘTf47%#4j+Α:<3\7o6,}PfZ/z/m%Zk>">Jm&8y ̄\lDJK#?H:a %a "f;^В4|?EOOڒR~tLfRؐ'H;M CO>kΨ@X˳0Pą0P; mav0L1^5ro/PY+ol253S:v5^RWU\p՛U1eKʐKJ҈h-riP2Ng9/1nwU&NF<屧n5\v*[CC߾uT|؋2{P`̣ %LF!g%鬈:8+u[x`Dw˷TyYvl?4/i>j뱁v L3 ǐO"f)Ң8p I)10hO>\9*>"i?u*RJ+YY!dͧczPfCwyMLr~ /Lh]1wFz?&|hI9=|7\HrNKeaJ}}ے{F0doLq8.j8u&)9`#´6lXeN>'HiW &ճpd"y1Bm6 r!.E[~/ )vEwW:|ar\J+x)(Dp9;L[~| eFy^JC2& '9Bt+hfBbiӚ hNd 7)bҢwTu#AP;ƙ^.xD%r\*%NBiUJd{;!P=ttz=hkó?ԟ9t>`x=8‹ =Q7KjmQJU-q (%*T X9Ujr%L5Y=?yxgrp3i8utC;ֺ.+ vr㷘y ]D!K#'s0224֘z<; n\}n e팆86ݧA! >]`A g ps|۫_L&O2}ޱ=@xjO8Zƶ@-Ed)iq \8[GKo-;G.=z'HR AB3.P%*r.1("' SoY(e4FIJYr1aIvR pF A&\rg92e1T|J|er4jLf{B;]T \"}Kg~;"idGcL&-1F֢I$X'Ԋr xBp\gI $Br: >qlINȭB#JBWy\SgͧEn4\bvvM_)&8k$iiQ}NtrURJ_)2ޕØRs B);_G^A|Ma~^}F[hFaAx<ܶiҒfc ޏf"aucz.i^D_vf5@JHXWX +C@,V[84oVP1᥵ N-CE?&~O4dQ+m A+!f٧r8 ӭS:8&ɜ=R6Dêe7 cnGoWm'Ҵ;A+/<>?4/.TZ_,8)-TTJ1]XyGA}-b֨m@gsҠcKaO0Ouf"(lĜ$f[WpGA֦d68ZaeƸeUiSvϮkZG~< Qt =8Бstt WK1=FS9)>ys!9nȽ1?q1D2K.Gxm`3;s@:dO]}fQ:%ll1xp`\;:5Iklˊ,{~)aɕ.6eeȩ K!Cm't$91},Euo^"Y^qoz@)sL*8˨2 C&ʠxT\@t0p4!FeSfǒר@8%I90]Z1s ]o$v6#Bh>Jҏ5!pJA-wS}DtZ&[3X(+7*oDiI8ouZ Pcz(ڳH  ;K]F U= WixQC)xC( ZKJh{dg44DrDy}s{t&+E).kv0Oyb*? ~^Lg 4NjB|z=-.oY^߽v005fÏ|HUG<_y=LJdj`1Xp[U :Ii$}|rMЭy]Ac¸Vӭmƴ 8x]U: Xm$ 'qLҤ!]%Ty~q-GbGs펻Olӽyv׽`an͖}jV[:\5}oRq]{pRj~,)-g[z Bǁs=I?ad;o:5oEiզ+ήܸm (ɭ<^xa췢1~4 9$fΪ(KBr  }qLVT7AIJoח}T߀y}dᙧd] .BAVYaɣX.}$)oIl嬠7z"!_TuZ^hZϞ6L6vȰ5eYYdw|pe“U_0=GB~x ?uN{:FIeQ:N y~R e?Ē~5Jugr剕WĨO3~@w&'S-џLmL(PbK~B(G@Rdb62>Zst,ig9ƚ_.ud22QhIidmU$VQ+Q`li=H2IXb_-B''I;dY._NB/o4EpV%*r;H42iwP XYo BRGPDEzNO X,fz[Y' y" ENRTReyړBʴgsY u3gIAMKg4K ,yPqŤy&DuFmu;-௫EOP,Ud?9AuF;Umj.g;$Z@\n>_swǾi%fjSf&դiWܕ)6͚yGmz7q\eYap3)}HSSj94OTW64*aHߟӃ^ ~]lgɸld- ozOh}AxR*ebKL/uS?^0aH>*x8Wn㒉;FJ.u۳zAYjum/Py11WH#,wF1[zJ~+v"MPpa2~b$5FDϰY*{"{~~PV*8#o8H.Cu_sRxCϷxL^l)3Ǫ. ~Zf$ Mq|w'PVe ~Z}" k%0"pPy)2 K%cjޗxy/> S+,;U`IIkDiD#8\RCe,I(v]ڥu8ߝ|iϾk|֙C!Yt[e]> fQ9F3rI2+qh'<ۉWRi5&Ւ]oGWeu^f/-iqY?L&x+ȫ񯹠Go[D'pcq=IEEpyRrŸ҃$U昋!<>i+]l TeҁFeD6dNR(K+3-ORp#UürⴏdyHLTS:Jg4HI3, 9%2Џj:%b6>MRYAߌVgs.{g+X팜 oQ{np?M^F#7VOcJQ3G*N&Qc* kڒ@Vt v@!I{I ahwR`&&t].+ro&i9-:M0]z_;COB0Nko~^M?Viwv6ZS  \P HpP<3dH2Ė1TP)Mc*vc01szR0H<.Ep:EEw?R]#cgGdΰ3 UXz,|D+W4[uj7 J7? ϫ7RNI,I "Q*# 9LpF">i]W,`Ps0g\頄W"jfΠc3I2 t؝xT}Aθc[Q`xC@BI''v37&8.d֚dh6輳6;S`DE5O&ФXQCFKYy}uv%m=.ޛiu! Ɛe ݄8FA(#hNTEg]ܗ!C|ysjBuI -IJ]YQ{:g*EKZy+؇&\]>7 6 c>1 ^T:*Op&`.xp\˟l}goчhtߕj%=)\ΧKZn8{˥<^}s1` Kϻ-㜥DņUքӶ ߋz$I}@O.>w 9!:V 3hY*3żM "#ZU-[#UAي#Z 1If,:PL@T}e`  6c>j%ҧl#Z)wx?^iuhl5]]wg!^Ҫqd{CAh}:2o(_\~V[M+oVD}l~U+}.v{~MQ~~7{}F>K۳|MѥBb&S F}O9kΆ?ki"eA< EPovztsh-bszm+Ң%=ºeVTimd0-&֙ HJs%]흒nHI7{L`m3Tj9h,h?2fIBd0='ZJhyHnc6RxTIVeY2@r)+IQܵSWW[\'}`y՛s+[ՈbsRHGTuћ^LBLULsѴ(ȫrxD9H1D{G> Ԍ\! \!V \}+)i_"\ijW@#]!`+V}+Қ^ \N"W@` UP v p"r,/\ToeSqe[}P^vfDA@ſ*b26M#,),)p6XJA3lۏ(n`jz>Jjv/*7~yn n'y|5 Პ|*ƣeN3h\MC5/ٔoKBQN'ü˸4YL\΀?GW Evwt)'FcmGh>_^O Z[x _(uU ]̈~X#|mԒaU+'Z2ÜL 'wZdP_GY%؝,JhI9pV\bGc`BOQ49 1Zè{`Wm b)? RB'=v+V}+R^ \*)!,kȡPR{zpglSnR![Ú# z|sZ-DCP:f|i)JR_ siUp% ITR>k6irBu;քC,q[vg uuWm6/߿9f_n6ڒ['yBY$%cȃf%H/x|i&ja^J1\w>߳zl_beu֌mS+UlF@g NF> BH.KetHC9Y`dK@4QЧ4 NRȅwL TLLZM14GKbP:#g\s4ݣl|t˪$RrJ8R`Yf7v+ɽ6+0j֊eY]3rDXb"Alb)0BeA'R8'Y$ZXg%1spx~QU5sֱ &n:d)80}AMBEYPQ'2Xt԰#6IKZ(vqݏ5Ca*Vkɢo Թg.61 C"ƘLf޳?*|=?ѴFJ+3AZF5ZE4XX` ]KMiL)$f9FڦĐḬZhX6+5F "IQpIQJp01rB.xe:f3[["6kW*̝^U:?y&Y篺Ubt\зO c -m:tռjDuj~b!] WtS @s㮡I)y 7,r\'6jҾ-~[.f֬o{4q{Wؘufo&BYO-SirP jt!}?oyv$,3U%?=u?xq>TP|m$`Ps7,`@Q 7x=cQ(sdp"fLTaP0j #v|Iϴp*]q5c8 m>oINv`=u]"VyYNr3Jɍ0]J'hژb2vQ|x}꣢nQE B^ߠUmm7ZP*mhXu>5!afO{jzCv{Sy.e$ Č=V8Í%\dv?:Ia5_8xfs7(~=5RѦglUڞ"\- > nA}TM.٬D8A(u) eʢ4IҚkHf1 oÖ́=_Qz㯷46"4rjALZmel-I >͂$9\NU^]KxbgWܞUmM{+Z+֠1WC0 cO$VgziQNmq ,Oԗ(S Jϙ+%LwW=إn w jdݧa^@^ڨ h5i+B2s^)8,; qND1\]f4LI83T{u XLꌜb`d_o[|㊚Q%qPaɨcoUZU㋧__dW[ؖkm}8>#fb9FxEe TB{ @ɧ| YYIۓ_֋%[j˔-bpv"h}WEiN޶}յ=cXO_oAQ>(FMXH2x@R CUS}GlҭptU8_ʺPSuqIߚa>N9WQm3wvShZ&i6e[R>Z+AwrR%.Iމ 4Y) f_EL{$x4,ŗN%u5~sb}v٬\;A2ޝmGJkE\RGSbc$8&0N̵ƍʌ|񥉰w\<+{UH[iT2ҍFWi:veu5䙥 Fo|OWW|7;2O;E粲vθgW;&+vv 7nyɜuil +ҽ*b: ]L-r*Gv+@އ).}ڃ حEdlrN*S<}dnz$y$Yx$t)mWBLvK%,W }BE@ksYa8]9y%xvW8=W쁮Wp< WEqɡ_K_A?1>nuCͲqN:ݴ9Pq?ͶCjPuan=z~l|U=o?ٞWHc3{YQU:F3,i!9oc-ca؏yhrlwɦd08Szd_^Z(Lr "/6`jsBy=X?.̓:^w@>ŽK{'\t 1y!Y%0dA2>$|paqywmvl~2,{R:PwTYkS*.ץ`㤔~Z` "iaxE΍GҍKd4ly [gapIg'伴eBQd䠝2&[Ŝ5ZTuCjͅG >!Z@VdiQ\"K?Db2.RM[Bq.oC|^K>PeXia*q݇BwYR Y~,ʀ_m~fa*XJbRFBAFq'6ȉ齊|kyZ1AZnlL0M*OSŃ*rv9,ɀ`6ֲP"H4FIGg.`Y\^2qD *[ 8>6<|Ale24.n㻕l^}8d'+_omx Ng 1A脵$X'CH78vpU589('KwrHoޒZ c A8YDQZ%)Β\t[:oK]p\@)Uy$GcD"q# Wt1(D&)f" K$YȬ$j"x4l.bV,cJαrM@֪.xm"YK&_l2߶ȤT[>"ɘ [i%pi.;Z-(6]1Nʼ|Z!Ȝ m;#LJzd1c !FɡNg[n8UHqTAiYҷʺK-:t/A :$ԋHQmn NX_J|^5 x)cE`0%?b*i20znMp!D֦?ƻ{O1t@"0iM92 D$ӟ1ODËv? 82ݽJѥ6!mG')ڀM\˒?:'QϮG۫S]?')EEj;\ Ҭxg4_\eA͢X\d_dV8AohIZ!s۴=U*II\ζt c|WY'aREa/3'q])NO m)j:rm(<]J^}Q}}Jm݋>[\ϣ.LEtu kBm&}ё檗| Kuobt-%vdvdN6qz58$Ut}c]m82 wXqwVg-hN=VBFt_"swo=nduW>FōO2=x-,=ڰ^c>z2viBWF{ekv R_կ}إ:rqhH4|C۽ >mq:{2R|o' σY~ubVm3@֕mx 7Q\h"ʁτu<^EC{tFmz,b9MVVwt,Ls|W[:WOnmO!a"8()d9br. l)<@DUBg ?^gȇ [?x:Cqr))h}J!B5’q&\aLYjIPX?rVn*Wf`̗"~y%UV;XIk뛶")ȦDYc8C LG .h:sX-cqRxAheI#^L?̮/mE.th_O?Z8_ґe#x-ieo^7C@$U1y%,qU$NyB&LV*nU<"apcxCXzQXCIB*'y$U0.%eK빱1$$#Q9E2P@x*}VÇa(tZ9\|2dIX`DckM eTR#g׳%9XHZWBYm唐Bˈ=묠S TVxj+H<CV2H(,%ϵfHɅ^ &06Enl<3;-jmMd f-9f~)R^辶>&&Qt$:9vXG(CV=Nnw? #l 5KX_빿]~6QZaZH@<Eww1!|kxU2 3ZD%5V,E"p+ZH-$Mُ,=##EoOx?_Zdpbh;x k}R5]s`nf {[Pv6hӠiwr ;9NNu`v''ԝt''Iwrҝt''u4e0'2Rndn- \nJI*yqgnG4VF ,)QYCn8m:TwdҹY^={̹\m_R*<,J ʻTJ͢bhRp*'8AŝAÑA$JFGEx&A) L"&N &pT(fVHu ȍAS؍/NhY_FzSZn ? rvQoS'. PSg=#e06_Fg>ɣV{DJu4`9,ey.0t jHUJ歏Ly/ ՚OZ0ҕO-R>uxa5]s]~+N1g{rIXH8bT`mD<@!tSxFANkCI"I4U s\Dvdp,I|kQynM덪 AJ@B2^CH#JikA(x+"%J |9۷+}g]c{JxVCI4G}0^va;9X5Xn*K@8Ľd:Xd]yD]>^kpϨQu)]jd]S z0{~6;^V6uOB?GG͚4ɴk+j9ҟ q=%)뾟\ӽ9jhUf.fGaOL,Ȍn̟G %C^!k5 qi7Σ۩7YMM'tkoȚl7lj%]g$H]y U}WǞ|1{l[ÌRJ\I0RBIC&K1O"@@xNf%ۍ>[4l1[8mڬ%KZ}ʹz\Gݥ +n5[mG Iڑ%;.NKwх{nHH\k 7LH3'o5/(//*=p]=w78O.OHM?Og6n8MTxGW=ͣ]g7Ӆ'so{tYA* &{^⽼s>)b޿iM"GGo=-kXmfBZտͭx ŮTY%69#C'm2}!jt  vA"g K4__ġ[xKkyo7hPîW+z<yg푗l4Mxq2N. gBL\rœ,4'_hl1XƒemWLou> ۅUvry}=@{=K b=R.AWjh`|.Ww./Q)&s*a ۃM3!|qJB?GT h]` -* 9׹K;\t&Osf-k|l.cメ 7x{7#_e>{3|.opG6zƗ\ Fk雵y70}2Sq=.|?YԧU'm'777|[׊֬ہJ+Z1ćmgi=D)g˯qDJD+Ǚ% _[]-% t&=I$ W{g0r"$͠:RRta"a4I F!%8ߟ >cZS|3ܠZ&GbK(See4+9P1Q޹uϦ.#rj08\}͊]soط\ su#fkwZluj@R&yE.z@﫺t'~EWGFçAVFR!&Gncǿ띲r&[㏇c)u rF7j*|?7LQhiiHrҙ*1ɜAաsZ:!Tჺ^Y6l4Cۏ~҂& V:훠[¯j83UqdڼTiEJS6_+nCk`{GzL븣CGZVp_`4(p\7Sz4_ю{Gloz#{ێf޺z0E?'l(]m;f8Fikqstu3$.=ZFk];/U7f7sQnH4iofNryV_kLwwJ4Y0ocE ܣjUњ{s?> q0kn99ȵr_zoz_8Yg+1xoה8y ZDt, m*eܠFiEj ) š2XOl9m~&m#p*hSoSYPc_BF.H"\rdN2t sΤVO˽዗ M߄z)æl]Ez'WqzA ]#e .u'Z?Ƅ(x{w_*9{SUh|Z0_ $(JHWqQ2TFfG)8@_cTqzY}PuOO|.^n*j1' HshkdXhXNz)1nEӿ Fi7('(w!'*rDB0(OH1U8;- [*%\OQ@ ˵Q.袰V0 RJ-<g JuZkp* h$(FPQ|$ 2O}q88}aE-aEm sW:–$K[PLN^z4?÷Cɼ@29k˪m%yZM0]Ӈ߿~EQpst-U@C"ҐBHΑSZ)C^{-IX1蔄Uܢ~ J#c1v#c9[cPBak.beBel\TMvyg7/y77Op+Gl DU27OG~kRF9@ BѠcEpΖs5O"Bٳ0g6S"3vDD.xЅ;ۍaż.&Ǣ6FmP`Jsk~qțb|0";D' @&hO$FTNEV=U"CqO4fGZ(8b}i:<`) k(GO }8;-nqq(pIbRr,.¸h:\pqă Oғ}%QjJA@ω`#JKIA3Mpx)xXL:CxCHI<\{4iZ{h8e4zI$3U1ZS8{ztU xt̸'e4*tWqP4s{j3K<.9e׌'S;VƶrqA%B%O2ؠ]><S JHƝtVO! iJ2F/Np"vV[8QȤ4K\CS0$S2tT\Ѝfe˜ |[ nSׇCdj!~k}v,z5;t$kVQ&2g:gDV c<:[1"AoM]jd'g"@r- 4V$mFtT\f @p]ϋ`X  E`_o7)kZHF4td pJY#2+hpi?O?}'O5Qi(xiA܀9@ •$O{`>RX 9 DBՊ+4qJ`21.I4H.wd 0)[ A/# nΚ mKInn4G/{xzrq,+d'dzQT"7]$?zo#M8 iڴjoMu4d{W.&KKn%.;_3;Wӣ-`fl5.U*WPm.F7ͩm-V`V@\N?o9Ocд.),ӡ |ÒnA>(CW4mV?{t`ڣXGXs/G$srM;RciTCs&zGTjꕳQJ^L]|o /V.9$ۉWMn~ D^cVPD7hjܼ.,X^8뭢nQo,QoXx}VcG?憦̹#}#l&gMJƻe9]ݝc {G,[ݦޑQ]2 Xّ bxi5 q"3f}凟;uh:㫁zVݸ8O> k Exj)ƍ7ְLő '0@4$ $%4(Iװ,bx2^ 7U-='t(YZ(rK,.EV @Yw. %-S^;6r|P[C/ jw̅]>d|ӵ2ѠKD: rOPBRLڲVu_) n f'}*ĎE/_8~wք &~Oß2ME1"Xw|w5S Y9O,jVvR9Tl`G$LIe-BEn2əW}mξuplM}Zҳe:nm5_.;wZFoy?շ/;g76|KLFY .K"jڎJnTTNCT1L̵2:-9-zqiEصe̶Pj{ԛOrZ?_R@?Z/'#Yv9/3Ywc+/;]qKޙ++{?hfհ;yBOul. }^W[N^Swsm*ꪎmVENxeb̟-pHc6JJi+ &V ')'{'M"0O3&zSN0Ĥm2TBL`;I%,W*2n!c() m;ӧlVmgIow^:ȵ cc,KȠ?!oiZWԺi]5wOntfCĕ_?uԲ]7_ywk ̬[4` y#tJwaRAh=oJmwmPYo[qx0ղgs;5M# -pGWWu ەp. LJmg~Ƹi?߀ dsO Z'3R?tB6Fi ~~zmB /#Ux)jDT7PtNe_8)7ui.cdp@^1>I)d#ƶzXH J+ :"g9+x":]@yW-؞7Ux8Т'ꭗuJ\PeXMBÓ1( !fQ&㲍&'Ve南9IH@2YLb4# نRg$j`T..?s{ifBbi  xPef "E̴SZ[~I*wC-Je2E<~0tt傗*_$^%RhQx*ΚvD$c$c%S=Y~!ko:e+%:|a80ĉ``r> RC/Tr8aA q%ffw^ST^S~d욎qնA?ƶjOh! oLS@W񂡷0cK#J8LH>$ƑgU~  梌ALIwL͵ yM0S'5Z`oݤa|\Λz0{)Z$- ȹR O"b1`:a0d,r1$(S4E6ZK9-CBs^kZZs! SdYkOȊ2mQ: Yxo ̢|_BZ_(0kyu) tgIË|%SJP%m1ţ@O%cRFB DHki[Q~2?OoZ՗^1T K^ c@e6r9j%k3hզVOi3C(QzXY |܈Ї@z~Lm}~:Ls}U)гhy$t{uke ZwDLX\PbI [JkO$hBsG\B+*OoR]id*[}4ꪐv-+V3{P)N4oB]KFgE ׵¥8^UJ #ޛ(UyJgemj8<{5D<+#ps7u]O%{p4˵F1Js%ed0?ŕcjya/vRt5.a=z^SEw!w`e̳u p;h:qNˠ+#"LD[I@iYX6\Q0 aӬkk\F-g&mTYJ,7\/tEC%Xf&1 3%0ZuIBvp ԁ` (-f1c &Kx-;iﳱ< (tg$B25;1vǍEQpI ¤0,,{c;&gو^ tĪM@k>؀ Mk4B:iˢ9a$+T%La&lmRi6X%Ix3Erh[S|'Ѝm]$IUfR ;( ,3AUbFHz De{!1ւM[ [ƮXd z࿇+ EISL1^%r,JxyE ap`R p#X("#;z!Ni,WQYjH)ZS{NŒ#5XV=ؤ}/A~Y7/Q40-l]΁?!]뜼oY%cP QѐzkDoZajl};k* Y[0m@Oil©:lti }CL0)Q32`p('SI|Nql)\{3 H&"ˡꘅG@VRh_Z$3S((H.Q4r/Op]ghvEC@L ##) 0@5+oz2|)n^.Ov<YlgVDpEk3Vyr5)#l]rt$TuZˣ0x:"Y+4};? x{/ReY: u@bX: u@bX: u@bX: u@bX: u@bX: u@bU (S``mx-ѱKEdX: u@bX: u@bX: u@bX: u@bX: u@bX:  OHD;dt@dWOEDֆu@deu@u@bX: u@bX: u@bX: u@bX: u@bX: u@bX:/UBhS)Ztt@.ឌ*u@RK}: ZOu@bX: u@bX: u@bX: u@bX: u@bX: u@b뀾Ї Ͳի-5׷ݥ~.z,O' dK F<̍Ȗ=zْ_,[dK>ҦJ]@%?(\;۽ vtz:͎uěCtt^V]MHQοWfMBsqb*-j^.OphYB[-mx?^%=ekR  p_f#mS`S7^utH8h8Vul1zWQ=mz(~LPOjB ~#)Yf $y݂cwQZՙXlk&+w._,;}m+Y~?DŽ$?u?'8 %i;w-wisr]d՝ >C9hpirAp jw;4ϻ%\К&s }j@M*%|97R+}46w ~Hmop]CNAQ wfhy8/$]76ivFU׮?t@-?`oNnL<<}ko>K/Y2/xӊNXC-/4]DG=FnO_mojsbY[>M Zr9Wm6ڤ oKO+S۟0V&3X )'R blN&jGXR@L6 dO:(RP!RJkqYy#*1b P$GWZ,5+4WU*Ql yآ*5$bGcǭ,߶cF~kku\|վta0e_CU=?7{=z󃇋>CފnginB_S1:UL%߰~;HyeQr?nzPFYp\}+HŶ?99I4!FC/Y07 }mo!ܞM֫Vɳ vn1rݷitOG'YiO ɧrkWȯ| T^wnIJrfӝL,9.4-j5] A%"4>!x}Ho3ew9k'n9tiڗڢ-砖;[8xfy}W\?7O N11ŎW ގlc׻=`;ܛ?U+eyzXlA[NpAwjR_<^c]׋nɐ_5ixϮ0 #F/ K CFc/_]уEo3 I\=LK8{*RV4Pmݯ7:za|fP;3H(*<@)K&TuV㠇?ĪZլrJf;j2 act?kĬ{k ~׈GFMFd^f{Qi'ZZH˪[uYe$-.*Kh_iGM&]d2Z.2̜'ꐡ MY{8h)ZN Ծ4EL ?NaB99\'vSώo7F|vvj {SjNOljFB|IA+mUUP'sn6|3 zbJYi=0g*sn+qc;<^.Oel|1{=pQ^+Y> $콭Rt6a'X:k>b!ד默B-.itX,agwb]o6={j*K;+) rB6 B TUStVM}o3؜`BJ3>aM40nn#\l4Fv7=Bb0􄌻Qs ;4 Z/ψw de&o}"{f]?mސO>΢@0[w$~E{[&r~ }og[E+dOIqS7- žF fgf/|1{uؗQ0~6{{,ҾV7g] vBmv>pmS&ߑmiېOy[)][jv7nǞ^=5/ \\soxy|ol'eL"LGw{Zi?nvnZ|9cP=niE7'6VwCUdDddSK{ALĭu9!J1Bn nITNΔ-I.G羝Ⱥ/}&;lkz؎vjyC,O-y_UӫLn;;dwsV;nj~&_\+p6 ΁+\N1gm :N@$8XeQe=l~z鄞y7=驑h*@ i9h CH#JirWu 8{\iuA8z/gYlQMӢCg?JԋҞg/LJxVoIU=7q8ӆdeT$w6D N,4BKO-~ûal_5&i:z_d#9bl9d#Ѧ%[Zmڔ׍nsj:3eϷ|32 2OC98'y0^?QCE⦈-Lt 5/v1`!6|sq86N.LI!&!|yFoưs"Ț&*gIBk? $8mEH.ܼݮ&ZvgV{ñ*=Kbv̱#tE97+p!MS,v ,>2GAY14{1ږ {hkÂ<ǒ9\҄.6-G ZFK;.^Evх'.HHBPHeBҘP>RxIzyV[lqR{Mu?oǃS>1 mpݍQM64年M=479ݝ.`g-P{YQFeGÎ^_WjkaRw]7$rttlw ƶ6$^,;e#kDt|?6>FW !5 t>ў*.)8goQkϤH`E ,|I2 dos{h#|%WɴD ` (3: 1q sGJBWtUq~J+-&Xr`4vUywۻ%>* >pgRGm;TSKg0ݥ3~ҙ[ ]:s?bwi- 'OUV>_ɢa#P UC_'p%NP2A>=R3i ڮJg>`๼_ 9)jhG9.:z^,ekE֏Vca*!hPB6&aJ sLzʈi7b{VrI>8>7Z4F1~{0@d.5ad&qW Bɺw*_CSL"UURwIl 75T.V uZ -``QKJnmkp=\S ƻ8Oq2uy|!fb06>~17Ln&ӇI6d_;Lb2&m<\bkmr1o{ެ-ַmoi!qs47Ndo' f0ѬL`:^q3xA,㸼jSiH^HS]uYbg|TBߢd4PU }õۇ7TԏiPߌ33j݈Nw^ U3g4Y) --"Ig$s^0Qj,V`w5&DFV NBe@kͬ +_->}<ueh:Y%y] O ƴtU:V;¼& ުOz> WGƈ4%A kp'7$s))\Zύ$)]Oj*8$=2I5Zks( JUyp$Z38}aE-~E]TŭfhF;{\ ~hx2=6zCI[h}v{(&4JroRs^vGyo?(.nQT_7 I9:jXDţs\k*(NKaА@Ȩ4s$o(C!D٤r,tJB *nQ?MR;R IƱX c W%2:),M~|\w?Q=[>:ng!*_6>ƽQiN|P4xrnIGO \W) };Mb"WcQD<һُn: k.湠v180jCڽSAcK%d_*r|/hd4I&ba ,E 2h# քHb!?h8"Fօ$TGxίbƩ_x6XJ>ED]u="xdBq4q'#S*ICHJ*!Xø'?h3-FSB4 D@q@MTxޣ'䦝<ˆXpqb,}NbRr,.¸hz\qqă Oғ}%QjJA@ω`#JKIA3Mqx.xXL:C$$Oa{zT,n4kh5hog?S Xǥjd5բ>k#|ø|"FJpgA3)S8#S_Xvq2Ց{T$ eK.B6IdA@}0)x;<7e4f)c$pw ,gՎsULU04CBi;f)WtX;l9 |򋃑ˏ[{Cl4McC cZ6({\tV$q5+eV"D=#T1PDH"r+wxFrå_u|:=[XV^*x*2D/Ё!ŠhP#6ˬ?@v̽TBmc#n1*AȨ$6@;2xKY#2+hpiJt"E LDK(p*bP44w:yb)S9_XVcA&>pbηyƒ&N#V&Мg"DcIyVPU, f83McC64m%|R0hR`XiPq$y&KJ=ͧF$EH&B('ٜCNF540M" q* :>%_1ݚD֦n!Mܫ)"&wj%|NQECا$Gs*ӇKLqHmY/,;|*x sR S/tV=![D.adǃ%oL^.~.bj> NħB#{u(.Fw/ݐ;rqsxƏۋ86UV$ZؑqzB1 2[UL+b7<ϭ(}8'F(Mm(p%6KEsE3Z.;x>pU+m-_'Y5.QshEr-cMYrʫNuƮ\1v[ޡJ=uHBnMUV#exѬgƫw,/k0:?]׈:AinXNSh%|#Y}ᶩ/2I7?f(?xpbC64e)A*JG!dCo4nF{'t'mR85hJ&-)g[mѬ.`^8eV.`s`Y-&Gfo4#dž>L,xv9`S>PF}0pafv(m3ʜ{s̥}sQ 2ne7W"*I#u&JW}qf+E | } ToUWu\fpD2$'ɀD \1 M\\)}QH\D .P8WBr$z8 f£atbgJS7_;>A2i\8/2\|5* 29`#9qJ %K9, XԄ8P1Qt;nxyGK{0xVF-myp&~Hw{TqcYpPUc ٺ`~*GD's']IgDPWj*5Q e=M. Q(vk%L:wAH]:c5.+ 8OJ yjvY[w+|N ,@>F`5d_q2Gu }ё-a?86` SditxEJI3JVG8K= WkU30J5]&w5,&ABH k-XD0= E)Be雦곫3poҟ>K4Q5qkodU]Ŀ',&XD\?a67vk6O@1q]6  a/CoE#9hk1ӪW?& qݫ/,F(i|Y׸L~ [5mΥT? G̭`4g4WJW݋ R$ +7 Ez29 uco.9]HHB #i }% QX\*i4sh]]&>={1M9Y-w~&:Ca{#[޳pa7>TG9 >Ⱦ斔|!B1 (i)Nix_|TJч娔D%В^HPiUo|z2GsgY`EmOOsNdO~+jH.Tg.?~4?Yѳ(ۺmZ[V)N=[Y#-Co)Oh%oo ,MƼmVB/d10dvXY.qGiܷ._CW lX?,^G|"X qo-.1UQeol~ 0VYaRE9n𴃟Oc0Fe@p aPBiwPѱa@<:(( #ʓV*6_"uǓC.oab=y>v>Qqn]{BTY"B޻֜)fJYQ T;QT=%t XBrY#yKmh[mMrіhs ^rJȭ&p)s6&L`ёpDmBgxZ2<6ttY YP;ӊpA\S y&&hp\r\CH!X)Q4ψޠs"uA8:oMϲo(}ТFg6CC:~q`~0^jÈg=';7L@@J$?}d"*wD,Ui4.&gwff.y{΁Or֜Rcq:dp%ZO`Et  zz-\}$PJ!p%(F0zr/~*z&Ǘwd}/MFvr3ZYw7~cI?~רY0&q2J45`qlkٛMMo+u--Bʮ+uK9FpʌcF[5J -2;G=8J[; y5[y^5PI=AFSyDy4'WMgM+<#?7c"Ț"䊀})Ъ5Fm:L7#K0hssu!5eR-=[[NcGY ڼ,šck8룪_{f.T+PD)ě[v$~ɉ)M :g'MMUs@94:k͜6kkÜ l)o}؟¤8\r MxZ5Bs(3wr-_p&ā)) V^@ni}ۚlƒڂSLe>:鴈 V]pÌ G)y^s<*em'wы{gRo#Ok`r"GKPoMrn42LtZ]1oqr68R`<$ON7cDgFBv{dn7ؙwx-K$'q_IɒcɲLT 3ILYݬz؆ߍ$נok,֟*2c.Bۛ0dpEh݄=0VY'k$hXgzV |i&X1klv\t;c+uwe7cvu}0'ǭrgA[<"Pg1h.{FmJäԥD# '763sTLLZ ˛VOs$[gKa:Z䲍'(D#?_MºTsǙ USdyt5eEw΢m"@Q62ꠓU)E +wpwXg%$˗t~PZ*o1+0 mƬ /e hJ5 ogӭNX)(i:+s7#nB :z挕>zÅ&1K9ad9h?OGZJ j432SeTc|b榙 ĂjM3-q+r !-i !!k%-sŲ!^)Iol@9ML2RzKrk֟6mM3˙ /ZXgHx;mwҫbpfdz+blSnvr}9ץ%?EOśv(2ǩ 6֯ NVUo}UQ,uU@avqR-T2+募|!;Ꚕ*r\{UI .VU-vYh_ c$ٻjۉrfSU"4I# ~woh@ϊ__3,e>čƁc;/ޱI4gnXe}W?f`  =F;M,~!sݠ=B9aǗT+)n:M~x}<2wql4$';r@H..skȳQi4w籋0\'їZ*;ze 6jl8E]F3_G5U@ d{6kB`}s]~ PM- w * J hFԨ:?'Ύ:\VQ!'F/K*2:* ֚T4;Ո6n["otMNl֊sLNl@J`^' , e S5֐bv'oOmMY͞/]: 6Bȩ&+D@9ي{ >͂$9,\ͭl6wʁ5с{ n.wgDŽ];f[Fς;Lޟy LG9ɗ ,Oԗ(S ̕o;Hܟg"$FEXEc\YJ+eJ`dܭk:UZM⛛/wο'Ls-Lж:kA3^Q:ˢNy3 YѧXXͳj[gaGe jUcO?.J䪝v mn]5[oy1,pm X *lh_CZ=ښ, ëӳ5Z\Vd4w/(csyU;Mphm,WhH.O'z3P(] $oLK 02(#%s"BXZi)xS>~"Yr$GI< oVf*5cWIX5mw/9 nE'9`dcw2L8N ^kbB5qurVt`_7÷'Gm4\Wrk˟ospYe4ilUKU\}}PRqo{A`~o 0Y;@9\7 R8iZVQc#,&fN@O &ʳR49 STTqS 62&Wi [ӌ]P½zXd]!3^sJS_NIwD^Ol8]? ǟg7RNILvƆ*R&e8 QkEIgݣUnx0 hK3MtP+5Qg툱 LPrHmZFĹxL̮vkڱ+jQ[3؍<ȖJ:I=!0bRϢ Jf9Mցj;ab[,v@$`!3cWyҀ51Xb DŽ[e$o5qʨM 0vҏ]ѴG7ib2 !DCĔҫI)R%2Fgk8j"5`xcui33QQ͓ ,i&4w|h;5qp-y.u%mm=.n: 1dHu7!#QjmJU%Oj }bW5 6pރpvO:DFn\g)Np8k+b{ Fi㏚_:!myiސ9 }mގGO5e}}e5aIxrHfG{x\\DžJ/QZU km9ozD8-s˧-p8}:[J!y+K2{fPp/2e\%{hqTe'Yڏh.$T@y N0IQA{i*D$TRd &K#ZڎhRx?\iw)6qYfiYo!76T%+/Mcfz b!#ǻyK?]wons4=Ը9Gj,QnޯZƛ-Z^s>L ͯn6no2=ix+¥^oHPynyMΚY(Wy0|8Le3P[+rOo\iS-Z0#@M.tJk#a1n=F꜑n:dN<D Z( 2e!Ln<$C~L1HgOS D]Y*u;]vPeL}}( fT')!}}X2 ~@p r8TP'\uPJ{zpeFA(]j4\VtPYb{p(H> #__޼*&Y}kgdIYITA;`jp-` L LUPJWz0m c`kr0p⚃1AZA;*QJ޳o薟#y~#AggW煫I\FäԲ[pEW]?=5Yv@p%8#W 崇WrN k@ZAI 1Փ뛽 (Pa|q10j@ H%S'ǕTcsQ*eȊ*P`9هѠ=_*eRJ$/Vgnw#ܗg~6?橨^m^Kϵ5?|%ҡG=8i\+*P}۠0+ѝO5$t%h/r up Ti?j=%UZͣUw~Ըl?q^OY/g%%jRݜ(s2pAKWh`jXg?˹d[kZ[3h|m5ˠTg16O-ۢj(.#BPZaNPJzĜJXkءHK:\\pJɩ> =0C+Vu\ &ƈC:f-%W(xEy<\ib _"\ʄ:$'.X>AU[p⪃+Vγ+N\/K']dghI1@q2_~E`jv6:J*u/n"Bp#SW#^ m3)9ٻFW>.&`02)YHߗR..,]hKb03HG&406"WK)xƹzjӮl+{Jg*U\}gծYnG\- JPjd /̊hU+G',!z9I95.s"Ȱf&ŭ2^{h!siHZnB gw9+_>-(/_8vمb_q\^[;\\1 +dXs}3*+存̭anKCEkޡbLR*n\es2z3*s嵛̥l{4WXE2WȰz3*+njz +[raA͘.[1W[I\e.Dk\ ?Uf͘+dAW[z%SzJ!WP()g0 VUV^\6v͕L[ #J1W]u3 rտ\2h;4Wd\]s*F:A3oktq\joL#1Ӛ+B-]Rޣ6PQ)E_.VBk%PjI4.~s^ړ{nl߭t>n~=3[N vV &$!xQMAwL$ԲywECrjg,&#WˇO6My'Oqu-ʇQ(Ǽ?,e) 4DvmѽQ)O|\WzIk2bx,?-dǒPOY~1v5@3 lSTCZ81E6D AG+z_3\ t;Eݑ,ޓdx@1'Ay8nDP?-l n2S "z mE6);.ƣARg(s;ay,E)o*c˯> T,߳&9eꪬ=>YoEΘ9{Ӏ0?8]CY^.BRѾWJ>_ZuA|DL0D& dF)SQN* #"P!UYm5j2dvE  a]iRAZMMaLb`LԒ" Ղ k5O&)S4 VMAW$?(&&KՂHȁyJUtidp~D\ D3ɰlRYg^..O֫R0|ձk1EAJq! AKViT/Cw),DhdhJ qcWc/s6z1԰;*”;Qw|gA{CvrK0: '.;o9V~GcǘxphZgdA~7}7y,E??zb>^G2%\$ԛS4;QH?sp43%Lh6VێA}.M~ lQkuwXp)#;6 `g=8=lV_6|EW)nSi88n; 3u`1Z*C7] |za-sY0q{ NG-Ol?w_uK?_~>_e߷7l h,mG j\~?[Ԙ*lrрNPݗ}&};v/pr1e`m8Mv=?ǿW/O[Y%[@U|jԖmRm-B %dq+ !ȭbW4sYv&AԔ#fhӋi wA(e)A;`9([z1=m7eݟ悷"*,"40D`4RVFHJshE%PSrL7{PvhhR[V[\.i%)ZcKT!ٯ[q5ihh2VnVZqXS,/Yf$o"Y:GVAZ}vSؚ6hۛѶ[Pk0X[umVH DZpGT i,Kش|.qRPHQri +u :㦅XXI7iΚNs(DdC^1w@I\Hǽᢘ5906%Pڋ6"M7>p>FRs~BhBY#AY9OZQm$\BAQ3CV0Uޢ8|i fЈzI[g¢ L"d4[el }7] &JL-8eƸeQFQ?&}?ݟBpH>? Ac.0~&PG1 ]Ucmw<[n.s8.pqGh;Ӎ;saQgy[ M Y:8m:h>sy%XϺXyޠ?x*fQ2MH4)8K87aéDadtmbʹ?G})]+Ns1rr~ÿ֡cɸ/oSeсC<#D0V(FљhPS$B8QrT!Zpey!ƭah  ֜+!SM2pZ4pʴO`|.Q@O }\D 9 kf J~\2Ի\M 2T*pD,Ui4.BDoX?ם|޵N9ᓲ(5#[iC(k?AE'#uAWo1ꞈ7ˬ"8eԔ%3ilT& ۿv>FqNZ]3!G2N;Yzٺ/SFͯ;Ѝ%LYnF8vyb%iN&y<A~VmJK˔q>6AQ+F:f'. vRj~I5([zV[NcGY zxYcp̩;s Eܔ V*ԝPFR_RlڕSH&IlNdHͻ fu>̓>[R!sh9shuZ9mv G sW=}sГ9|Yp- [ u%mGsZFNk௢K5Sv$ We/ -6LH O?_RhnvugUv |9<<9"5:ǰ{;K1Ӱ?+V-YqpRv֒ -%??\6Ay@Q{ug1SżOb%د;&ZFc$f#Xmft|?5xrK.7CjؕўJ+lrFH'@6D|Z{&bXPtQ z#BWiφ2z|%wDסU3)t|@S?R5z,>=8{5mL66h쪶#>:4 VpÌR.m6υjh4o/.Ϲ-4 0RjAY(8 0R@a6Z^~we NJR4QnA}rFdt甇[,҅\KX.Fib_Hr# $=2wQj-n< U8J)z?\i*G=紻Z @P.%,((W@3ʟ~Fձ8qŵ 9Y0E}mio|߸?Th֜iz80Ԥqӷ'xyG%oBrfVOv`pbRB:i`i`q,S$dy{n+hEnjEK`)Z` 0K ̝x'3,oLr @Űfs[Ӊ7pvL8t˹.ќQpYp0?2A9?6 Go[Dm']ע!ޗT^7Ã8#L֎$#~ύ2NiVUdTFHSГ F",MU.H02UZ4P,T-cw=/8V$;"/a9Y@&b#DSSw!*]e!ʼINBZBhR'kht{.2 a+JDll;blfԪR&A5qFl? +ݚvںej ^dK%TY4Aɬ5:\myg#L }xh,d`T`*O&&K 0v8$Qmp~2&xX+  b[q(" 8 }B 7C&)1*y`ERJTYZE״UDLoN5mG-XF#`&*y2!%̈́NSm&w?%jXpִ֥P\-pq{9m@b YAMHhcZR:DUte0Cp1 ^T_K(kGX0o<1?b|ky7=>Nw;7|qwrU#Z]n}8&0)SOw˭P5y\UUp6X2TA=rO-p8m:[J!y+K2{fPip2\%Gh9ǎt A"\I2cԁ`+TH٥Q+,}vhi;K:Msݥmf}f޿goW,*^y_5zn#QJPj?.W9]Iґ=7MUx36hni |*W.WgL7u6rr⦩FI*eChP&W`:#"{" D:?t-A=vir䀚Z`k\˜"gRpeI;%&ay Bf2ƃGm`@ U;[5xu1eb[&iܻP^g;Chaó_ᕍ'¢RbuwUc~?Q5RF04]D[ZjPh(B*3T,UR${ uw`S FLiYW?+>.6.f),͝xLO>IFS!T Z}TiћzTiKKZշW+|J!(Y0kJP2`U04Y{w?~S\#ƉTA.6ho ty9.Q96致|wKm{ݚjmʹf;VlY>z~}?T V\ʾP5S5zj e=+XRBq \5: W(y1+9E 07pji]+R^!\)=+ 7pۣu@JJW`91=+Qc \ J)Wk#oV WM~.bQRo6@*g cDŽߜ|]g˺hޜ4ߑ% |,ĘT Xͷo7&Ӟ14+X_`JioW F=٤eR {pg֎M1h]Mh<|?iڙ9l< 7./RJ%"r9'au@V'5L0OѲyՇEroi,6 M:4 C% 6\Mƭ/LU۷{h(3QncF-iF5XZ-EaN&˄Odq2 QykX,6sK*Ĩ$c`B'2>idYEo&>kTG&6덉>ѓLlV;&6J9\|A&sňXX^~GQ^|𗅫GIkxzc7p%:t5L \7p2NW(R\Bw|C'T}+ȮH)• +#h"\/8ix=PJ=$ќ~EBq \ 4dWWRjX ~Tli_,FqRʁ]F\@\AzՎ2uB)Ű0rV ?[(.}+VJyW%v5C jD ib% -cD2N!d-FquOI0vL90-z9܌ζ7Edɼ| d>A@͗8:V"Y^Id=dv[]Rˇ粪}j`%ir柤6=]l87WnSn$\Apz\ȣJ2ǕvJ3[ø+Qg+t\ >pe,'F+QiL WW"0ZGj\Zk+Q]$ p</Z•+J/:@\yib< DWv_72LS܂U*FNW2,qV+= Z;"X} Nx{rNҞi0 %4qRI C̳"cR YX20+31hܖkk.kq)|KsΤVkO,<~;٬ݕ4KDO@)6{3N-rb0Ee\vRleS_$iߏtM|f}]grnי\WOmzB~my \AKkQU./Jhbd%[y{u}Wxj ֟^z}a$pqw|{_Ggx,nZAfuOՊ[rmN!_U–Z"N(ձZQۣ{G DagGW[= ǯ 8sk6mU=ӊ3Ƣ*|GhQHo:/G=SBnWү]_J J.>r[ՀۏWŎ.o9,O|jwo(wW|/G|wXgn^zgu_w㯯svͶ[컣 g恬Z5kx&j}Uk0,VplHa8"WQp%jE\肫7Hh@` +~v! /h^<μ=^tFeFQ܂f5eF0A+Q[{ *#- qe՜d#g&i:5p'&nN54̷-Բ,{ O@S "n Ԃȍ~e=̢2L-Mϧzq5Iu4Ijii*6xxSl4v$\A< Wڗv "ة@ r jM *]W+#[=p%WK4+Q# W"(DY!y +qvR< V *-W+Oʿ;ĕun\Z6sTjL"B6Ɓp*@F0{\A Wl5m"2G щ+i$ frv=sATҢևSZTFP)ٍu(5WoWq˦#/N;׸&4WTƙM &W"xk\N(ZCRדBt}1%N總Фdg Ϥ`eI}nمg ZlD޹+mHeg;<Ԃ̄}m:IuofjA^S3Q U\MitMTksy0 7Zmc\=izG}{H߿Ў(c/. @O(j]-o9[uOT}qmP9> l>;JK.jeղcks!Ꜳ%nF՘ )cTxv!\uFw+TgHx$ lkfV7j@@R+*2VD; Q Fu. ^Ls%lV"- ͯ.(qxGj]#}"/w_YqTHgP^<:fDy%cNs6 ~IņԼss^U(Z8U7UrҊ  :HwunE"NԌMܢ+8u ~B]ko6,RjRDZUcEb3s65_v1UŢ?'qR [UCb*CUW6쩧nBznJ!(R>Սb *`5K1:*D{j$Cm]jq;0͒V)`x&Xh6eg} .ڐ1(<ݫ xWZ a|KΛq6*v 'jͭ:<ĪWJŠʳ-cMmL%o;8.TEohݣkWgm˦q`d lB7U.`Pmzܦ>TuA.Bkjʦ*JeDԓ.eX|/9RlG=[ʮtW2 dAYe YdC4aiLY XḠ ꛢDd8MJ􊫩|c9t`Q,d[GoZpB]8B߁ [ㇳq3FnoH ib  ("2]i|[V2yuFޚ,FiC5( `16J2!B`9˶% bAvL_`dUV.HSQ)G!@7Ǡ1nXd5`wZV;"J1< e֑V8 iƭ+1x#1j{ h}j$w̠9*9kMGl04Ph)ѻ` 0#b $=( :jn!`Y+cŝ[ }FФFy&#aPer^C-fĥ*ʌY瓌1QTkDDL%kœ=T&^8)onLkw( >n|-^VyHMj'Q-7K9u.}*ʬdbmt%9T*#l`S`1&b]@ew*)tF\xo4xA"$Ӫȫ z҇`,gm[CH7m[|OC $!T_ )pHc-#q /}M,S' ;1kՐ-"VjVpaېMR,V,D!v*"}Xaݏw7y?YcBSuF)VXBhwpRns؋9HbRD b =QX X)LERi0%ZlJRJcEX``A !+WBpZ1mN=KZ,Z =[/Fs2RLGJM,6:zI035YREDi%&Xd{P*Z;z+rV5}E,2;a̙J cF8P#Y+ !HVDkq?tXҝ,I* kFЀ̬$1Rj;9VUrDL(2Z;Mw,aU&4d_l (a+)0jN6vbxn I~.NVioί޶zL`]@wX7ם6wmH4pKK|0]dgMfc $;"Kg~Uݒ,'vbJ8b?$zf|dނ> K˿{Bizj34kZ 8:XѠS 1e4 ˻{C{o~P7A@xp/arHEFhNKv!f.#T;!0 UЁJD_, wB\p=j;a0M O¼"`1\̃ ,@NY<.:t+gbaB;B`X!R*+ q#>YϢ\Ce`~ еaIA?z SZ)#܊=5܁ctgR]'LA p&}H+0q1?]*X~wa?S-a:aPn5J>׉JBu`@"m@h2p(?K ZV2AhC3|Q+}!VSqtX#;c|p*:v/t@OCht@"HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"NU0f[ӜeV:Agt@"HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"Ht: Ι4G: kot@WǮB^u@HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"HD: t@:UP{.G C:S#HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"HD: Tt@'ֻ뀴gų7SL5Zy7~NPsj"V#$ $[2oْȖЉG/[^ldKt0q#^4}+Dݱ] ]An\+luDʾU#.`DW'HWmy s%zCWMO :EVyzDWj&CWWG "JN0,t ,#ZKc~te4]`EWW&D;]!JG)ҕSkIe@3ڜ0hm%ʯfGŋ<6ǤU0=rd ~7_ymohx}()>ICw9]!\BW* CW]K~Gt/R''CTf{Tȼ+]qv=_Ev}+Dk!: {DWd+D+ͱԊ 7]`~|G5 Q:AtutzPc Q]]n!a/G•/th߻BO,J}+*SNWRwut+M_ ~ Rwut0Xy)d25p'ȕwfQog̎WZ'mEYkE}h|GxNa4 @#4/tol .)>ϛ$ׯ6oNWJQi_Kͻ=Ul_%PۃJiDk|anlm}zdA z?) Y mtH $#Z#}iQRZhiAܳuɞ >u/VBCwwt%+AtЮ{zDW>o{CWCT~(':ARk{DW#SZˎ'JB"E -Cyli1DW߅Zi ,CѕRV;]J% p7}Ц7tp ]ZԱ "]ʹ]`Do B+Di- ҕ2GtO"µ bGv(':ArJ+-H19r$p&@VwI64>BDǫD7;Yn:Sig6&Uhpۚ24ڳk j:$PY++d3sU!EH&;@4orZN݅Eg uczbxeٍ;Tra7ž}o߂--ߣ8/WfȪLUdҲ3>>]%mY޿7qZ̪}Lo>9@>Ff$_ W׳tuxr|/PrЈ숉 jH\w7A}&fS?Vu=_~;vxi[L\h8u9Ho*)ƣʡֲ)$W%SVD!Fi!+]Ee[p5},` M.S⼼TV 7ܜ͈(S;!g.KRYh8%+磪 4eTz^'ϒ fO kw5 ;qM_\l oXt\St v7m~u5uKY~Fuk;̗"/{.Way-)c S |mn4^ nG_/2,ٻrİIg[ٺ/9f uhag`\'m&]JT,b+|Zʔ2`*:QɪYٚ-6&ki& kNXmmWA+ڨ\6Rʈxl ]7C&dB`+BJ༮R e!,T$zL#Z:{ CF^K2'1T>);+ n&fEpI箎,ʸXyb< {:YXY2^ڱ3xt9)jVutYmcys0Zo90R: Y/l/ve|d+a^ċ o\OWh.iNho/?+֨bW[`g'](ayv;!O|6jg:wF{yxe|^8(Mr-@\nD>_{V5tn[Yՙ/spbgE>Ul׼6%HxeaY JXΊwyY~kWrػ:-EH@<_#heyv\0,YRrcjxɇo5;ׇ[67i&2΋F-=dr.|h댻1$SgI,Q pw;<%9FJ. 4M&Wf O|l3,QG_ƯJΘ-t 9M<]v+lè;p=T^W&;}ъmW o)3pވL]jL]^kBK}tX-ϡ5@+a& F ܌ B18phí6W0u\ kˎD{m0~*΁u`y_jٶخԴ_H=gw!諎u|{9o:oʞkkXb&F đ ;hK8HVIe]5m=c$^ F{j{=bpaoqwW펷 Y<⁋(O{dPX ۴ǪfVDU9LYlnr5;lFn}21]'y]{Oac9?IY:%mkGpU7'nsoZsOB^B9iTw:-W`s. dj= V,zK1*^O$vy $g ni\_FDݷgl.ݙW^ 5lk:].eq[~'>]sPF|Ŭ'S[^.~:D~A̼*3T4 e$_, ޚpw=俸G~M{"ߞAq-mæJxȍ]ۓb`rV*$iDUH{ǥq04N|_n>rG,FO\ bJtpZCs&.h-;1*hD 䋋\\nmNrڋԘ  A$uN# 6I=S sgC6Z.nEEہ|;cndcxop[n@׷~ɻ8Zږ⸞=L5{l?^)|Eطj:! %rQ;mk=\T+HZmzJƌ(NVBaMb.Um8ۑW醅fƾX;cXQ3^sJsq~Jym&&/;Gl r!('9Rt@`Caٱߔꉇ^W}';$@p^uoi׳Scw"l!s};k"*SRdElV ZF`c`+%/M,oic\K/ӧ._<1KLNڵ -o\pv[~}Out kc. ΒvwJ%]=w\w[.Za$p#Sh'kٮK+T~l [<@VzvGg?(ݒA8K.ɽ`c2[Eey*ηM#%#ZUr=؁5-.e.Ъ6ArN΀@W:st56VR;r77E^wrgۦ\,[^»wo{ 7mv?y&5"G}aȟ\{6AZz~pzs8-.lqsECi9Hˇݻx{/wyp{}{e<oSt_f5d]oHtDkrqDMKV>,sn둴>ԑlx/Sx/4? X`WV5|>h*r|b8AbPmɬh;yu}jɝ:Ouoڶ:`b3bRQ&%HNEb}*bV}E |} цH.ZU.Q5_qI`΢pvH˻7דm`Oleel#r,/a&b^lpy7bdR{TKmW#.g l@P rHx}-ѨU8c+݆.dӱcBt2 b,^'U(_,bY>5b MUUK)_2C~:*_>XsQ dY˚ >:*)*`)h *+eHa(fߦyS1!BE( jYsV6ڥb^Q[YRHVV'Yhc'dc%~i]{~m AőN/u]|.|`ι솻ۛI}A5PM1 $iw#_\*SGqn7rsM)YDrU,J']-)F rvBXDHcnd\}]fLN.32kR:RP*Peuϑ){s{y1o{~â4}b˿ʶLPBG l43 !"DlQ^4TSJdWoP )E_n_t*%,r+MKjI:c {ĨjVKwv ~gҳuko־F)3>Uz]1*#)KAd]+G%2P̪k- N{8FQ=Hz:"2Zmʹ=Fs2}c#ȴĴpv($tm:!d{@Z*2o3ņ%z<OHh8## u%k7S2@A5 B|NC+җ)0>ؗI ɡ^!1?X)ޠR%V=AP\h|X|V)!<;`+\~OfU/oc_j6 [c Ž1V'I{- 2mFD8"#0p`@t.PQYgNkrr1>g 6;NզD[\XK)1<)6+<&/u|osYT[3":,{e^SCTQ&-D$SXU0نe%X#tۻKg-ސT3$2|ų(ͅ(b]T&I2GUW&'*dž "GF7o2 E%9LظMAztxؑ|,Fq2F?7b?zF E`Rohcra]øfCMPrVI% >XRX9$<@8[DmZy=`|] אPkuL6|$CP6j3H%$Ct-#v"{BW?7OR|p8Iwsg}w}s|Yxu5z{/?}mOS.Wkm#9__ `$<  ܭaS"ʶ_̐"i9miVTWׯW"6]Wʫ$Na^^u2X\ExOa Nm[R>hUеQnp㔼X\O׸[gF1#:6~fBFf$g槩_LaR%lnX *jn͸e FH)6_<<\w'{n]ȋɓ.%E f6 bM>TAhBp]Fn$ Fmn[H͟RHfK)ݚ;M`cOY ܽ,clv̱li&)s\j*!ަIUS"1şG!EnN;L$ 4j=Y/gm~eE40;xi?zoSz,yNOpAm%$FM[jʞDѕtځBfJ _o<+^5EL$l8M,OH",5chQ3Hn0vG7LOM47ǜNtbxbdTdU5 b0w[ Ŵ"csg7"h)_&&r8{-\]_x`# RFa4֔cj\0por 0T($.^eX͞ycƆ7uƮuN1UcuhflT}*vA:Aۻ1\^[5 QX . +*H -ЂPe`Bh$.S8=/Rb{gl.VX+x./[d Y4 \@O#ۉ;˕~3a!^+k)L@yM>jb w`UTsc%b -V$O0\jܠ-)P^B"l)e]͎ɹQ?v띑d[:i0T m!X˜F;wRGqB-Y9HM0!S&1jvek ۛGg}V)Q*q2@4ܵf^^^nPd|tlP4wHv9GB25.ЁG^:c ~0D-FsJ"kn|W>[RurD@ TSfY/wZI%PVAyZç[gcc%P av] S 6IoS#4*ʌT Al=ē~JBxVu8^_DNA$G,˔rjA4DTv*:d\` JN"\D> Hjb Y!8`}+=S((bϹuiw/;G00~g$/oύMQO爄<0aOKDj)nKو$q* %Makw6gq6Uw682{,G7est`±.dDMȂС'xxU'뎼Rv~Z/рmP㠰B`D$4}te#[g )9R OqV EBh e% t|z^G7EΘºUZ)[OR 2ZɤFĈ,z/c4/B-X_Yoc}% ͛P-)C/VoX;߼E+|Ӕ,{ڴL G MU \s`Z AE:: y}Ք?Ext~4:9 Ó%=-91Ժq.̜gU)GjU;+U*>H_p6vf&j[ꇼz,pmek `'tCk8hV~]&5Bz4YN|Q˫YuOZaw9MuL*vu9iO+V<4v.sW)} Z8̸*V?d2aEE bX,<cz?i2Y8L9DdyL%ˤkRF#T Df#{+NЬ(W`B<UX-Ә~7豉hj4BZ"2lt<ޫ캙06j)fl`)K1O3V*VV8/9ݚ䰀 8aH0pR ˽)yfcBs@a*y>|u&}-Ul^}kL.*pДyI` RU\v_J)Z2~$RL@ɤěUd&ʭsX3Ulq.uu{=/erJS_NI"[_(x<0\cs S.0AWEWKy)h(U4 5yVcH`x( aJڤB:`^el"*am1Ekgc̬e{ văc-B?>@3QJ6\uX=, mV}" d 1K4H5> 'H !n!c}9An+'lև_M$x61|Ufzk}/10nBֺB BDnE`q"ф:d%lY5VR?0cu#D%#1S V i$53G˝`wj z\v8;'_g6)9V/zQz׋{`x"w(+ C01!䄅7"8R-)lq>CrGm;٩&my֪Zw냵_&X"dɨ&8,P_}yLGǸ. KlE)!]Xnu OL*hȧdS9=uLG|ɭgҗd$*8ʝ7\ (JqV{1gz4‚S9ڌ { SZEn_ޅCp= c_iybL4{u2XW~=! S}{~۫?'7IuxO{g%qI96'Y=⮋'cvs}L'-l0rnBUld'ި؀?.&o'C ]]ԣ9)]9fd?͊ ,`(;}3o;P^_z˫=wN1˳!~pӃq`])/v-Rf1yr{&D¯tBn=6!lV敹m@_UƙBK2t=*QD ݈BkO(_UW D ]9Hgtu FkIE aM0N˔Se *GHr0M5X& l19B[3o->Bԃʁ`<)VXKMhj'VӀUj]1̆l6v:6ͭql>x6-b}|(l=d>reRjwj8$.$5K%W[AZP,c"m:K6Qy/j6VW+9W+??^:~MZOnok*ws/#\ه:5W%}YgM!GGb*Q#K!c!c(84η7r劎&1RTf]M,ނ"Qv*dM5L(:X\nr+6RkJ=Fwۼ 8 Ъ V jL)B6 4"+4R,B80XtAlI- Rդ=85x-:%=~eZ/\"͞Ν~qSim?员"+w2 Ĭ3ΊsZ9~$Bzo=8}!"cȮ{1g&3yَHHHHy}{Vج8;tB['Dq6pֻoLjzcO`pz$usq 8-6 R}Ch;z/%aޠ?qF P Y^E ޒ3@1YPCK Ge?D|Szcr77bzxkogݥmⷷ/kjie <уtlfK!DY(ĘB!ϦNE-A,B[ˋe<00`HؙU5 0BT1xVCH0G7&17=/3Aj&p2C<̢eDr[Lpn'')D.UIvI0 dcU((dTe㙋JlUhx2<7wWy4=gsQ@)%J1ŜHtܰ>1l%.8F7NOw(\YE? mEhhfw (!KC~BNtܿs3G&_S\R[УvDjALx,4h_a2=-WK^c|ً+C (D?~ѿ&/"W]6|3}Ӣo bZVGԤve:Ao#岅_W 65Ozp^0ſ\ߕ-hy"a"П~D+]q_0l& nos2݆[=㒱{[9}gUqs{;-N((@ B@-<}!6i!=)[% o$ T=TnB f} A}[ȟ#/*w `;]?^ F/6Dh^q̾˧ztt kR($9'h j5:M`J)CULJ =08g'1D/bhf| #_O"[sNTl|$VC}Y+;\=ʹc(Ir-l&K|K}U?O?Flo%H֩)NRM>bAMqj!P +d0z*Y%9pr 8‰uͤXWgG35t4E)YYDiV%媨E :(RɇudI1 Bݢ)6eg+Tt&>}\KE1I˞%B "<R´U+N9nUӒz5_4:vJ:&CH$Z栔^ ڲ^Rt(RC< ]I-hD~е_[Vnp#}gȫ 4jiy]oդ~l 6Bm/Zlww;e:U(YrGmBB`ǰ}yC^8a-i48ᚁ=s:X㰗sY.boJoII\Ln>_^kpX\]\,<九]%}'ڽɾ,:{Yxy%3_'?˧K>-mmKq ґxb(B<^HL9L 8i`hC UE)[ƞ_j/{ra-Z@a-Ő `UWr*ٔ8CuF& (O5U^G2ANdUcIIjWm- BCgl&o b% }J9{$:63y1: hl؟yCXc𡻤\[ʺ'`rǟm+w6 J;P)EO"5 6 k.r(SEĖ{Lΰ(UMNBJg2 E=;[b(7YYZA(}SYy.害)^ 1ãFrMYVSUMt9ΆLV -e,cx%\t {)I]) >Ǎ=Vp[x㋒'_cx$x~8ɱo#9q`0o.uC/|Y%pE # ٷBAQ}3TrEF5O,jM8&%; E iP2̵lh M=8W{ڛN!7Vyњ-qL/9mul察 qDGGBGA+(Q&ٻF) kkTZZ>Bj ntm4Js'g=~Ve{PތRjӄDT9 I ( A$JFG})⿚CLy8yw7&l~'ukyfsN8C}F`FјMĨFk"y X zI#I!fsz`rjHEJ歏Ly/ 2)t{K-9>[+z+T`mD<@!Sx&هCO'֐G,Hy@E`#!:9G#1'v=A[?|Qd!H  #ǫ@t)8T"x$V |:3b7(o%WZdDi;;z/&gYluբCf?J(Yڷ')E-ۿ@ ,/,I8{0N 5'CL(g.+z\H9M;.a0Uug#m(HHVFE_6D N,4.cYXYz]mUap7E|00oQ_>ҋ'jG rsǫ1-.$)MԐ⸉pҾqn[yj;B̓i%'2L4"8̷18d%q>nACg!Cӭs|M;6;p9bn$;dK}stIEvIBq{[vVbgѥ]tn~6>4R!W&wwM w?^JHBPfBҘPbQS{;)[UȎrH]zFnԧ #N5 C=w1x8O3Hmͺ+jͼj= g7u6\s;]0SyA/ɏ=:6a{pQD lÎfr;aCɼSօbDE^c"-4}R5J)F;T 5xɇgC;_0!m޻QI]Y4y{8^Ix wGf+}ZZoU 0@s9 1q 󋹔Ae麪مnbF&E5wN #,$H@hT}TTs,(+ˍ }_!1jzH"] T09 u=Y`Ѯ; skؚ ĭe^l߯?~,ܶBY )H)C21)y\s#"0)WhxC >ގ$ϜNk,ϟ2VmBۛ0/cU]l}9:Ϊ|N 0TsW kTelHUp<zk9qL5FZ~m\u]Z@/w_A ?Yaާ s]hJj`$aCf8.D^ MnPzgxiBkm;=Mo_Fz_+sSew]ӵ>?l{M4-& ?P.~Z}+lń#vҕW3[)JO+[}m?% w&=I$ W{gQ0r"$d:RRE"9-(ܳxǰЪ_Y嗋]dܠX&;IGbKȈSye48P1V{{#-Cڻi>tr~_5o!Ů7[޾y!F0ȩaT6k;jI- *Z^^i4JD+va5+B.$2^j-mPDu9]Dshk&baI8%S W1LL451'(w!\` QYíNcXl9F(}iuiha6ʅ]֢RJt53TBI:R58s& (B HGZ,i)DMRr#c9]cPBc, f_l܎"$dڷ}o|@eGɧѰqĖ@[%s D'ߘ2*͉O.s y س0=WpJ`&@Bێhȹ*]-g?b8ŸXԆ¨ =jNG=-!:BP?!񳠽 h,Z&AΚA+I*I*BF͐ =X"EA1 ?.$G:&8?-g?6FEdƳARq,"ˆ{Dq+&s>RI:jERJT =q PA("M Q p武Ih 3.8J-i& 9/ #b!v9:qɱh q]J@>IOXF1^(-'(-%4uSa18 ነxd<<v]\ޯuO^k]5H e=%ɃZ)eV"D=#TФsIDnGk$!9ܟ^v{<6 {yE^ZHF4h`Go0K8zDf|c2=Mgaa{nRG!50-/ 90gA@BەOxm`9RXq, 9 DB§sO,ioc\$h!\vfRE:a_zP3&с MwF g83,~3(Uz23y_SӎϚ8h=͎j9 7zRMW ,8P)9ZA2Ô с#֐BW޾PnFz~?GK3mR85J&,)g[m>.`^8eV.zgsgY5B+GF8+i)Ǻ>}Xq9`L<0Idb(.Ķ6 4uQ%+(rni䉩tJL怵龺қ9cen5TQ @ (䃕T]AJ;D()Q)ZB-W̿2/ l>᮫&>!8o/kczv\۶Z)W^HܨGU/痼̏%4'?Ŷ`pC»o#Co?'܊x;>\6ʌj@*:UI%FU1Ėiz#DF>s#]]mo[+B>bsl9|3`wnqmq?l䐉7Hd}G/Kc[JѠDQsf>3/ntN*mUX{2 uoKmSv5. 5"T2N%HOjA"d:.{S J)HPSL™x-E&Jen'}p̋Hye=7)WLV'kIzenOMS2vASfz̞>_E|uty}.4a8ύEOޓijZƐMRM94dҒsG!ĞtV+j+Hv{"IPLM=3P`PmY2>aKGEP%FsQ"XYP20*m…Pv&ΖO&1f<}7A9~Zzyi>_7[?s(Ee kM`YvlV[VPr5m%gɡ8"X,$ )G'=0Jj6$%,`#Լda*؈;n#^y}\cSu]Q=\(Yt~.uQ,仢AKuEUVizpϾyᏓ+m:d8D"g?LFLopF71૬RX}qW?^g_ڰ4S}Zx 5]0yv[Wytޞ19^f75x|?CS񸋹=h_py@,S| ';>haN< hmtu]y@oA[4Z6y9?9"{@o4ZP?zBE(b*MeP!wN. ꋻM2(:]d&R|WsZ aGt~crW`rX\CɫJ9yUJ}N-gړҮvc8ey AIc4fF%{5x/Ã}X\PvV݇D>[z<Wў %M'Z|N+\aWzNh@p8;\Uq>biWUJ)zJ? ,`JK@p=UM]]UMWlä\o^]cozj>Og:b bJM MMK! 2ɱ<{[9Ck!_ۗ41/_sEn W/DѤbA^ n>!deXH?9,` Pt*¯!ۛs~|~ˋ0C{&*]. 4Nn/_>e~O|_)\~Wr[)R:ov:]7%l* 9eY[ ƪ(\!XbV$%w V%/Rګ? $y Wq0}WCz’N v5]Cr RFb-L.ulsՍj&5MF}Vڙحt^fnS/~P%,:EKXDl(bN3mIwd}Zc0Xml!*1Te@2TVJ%L 6' 52v&W ;ӌ]tGk$.>VfZ- O?Op4a85<}O rTV CW9|*u/u6ND"#ˤJ6:ElTA $bOX d0$+=¾pN:MId]Ljݙ8w#vӊ/ݙvڶcԶ=j v3aht`KQ6YkZתX 1xbÀ;â L4 { ًN*["%đ1_&ة&8?;njz&6ll^#']! AN! tjM8NO0w*MDB s(ZM\$H %ke 5:&R! ϩk춅3H)H *:&X1R:R{ P v&ΖQK4ݣ։Jif%JrRϊyhx{ņFd͉]s[>򦳜rWR$H0>˜M6M!*PGR9ފWt_X?mIMڄ*CIZ فl킌:yg,[Ft(]-ûpVjƚZ測&ƚ!$K2༎Bv$d_sRJbP YԿfT[6NI{ V)kѳ*ϪP}Wp%`qT㵭 iEd‰T[ۇzNJјD3ƙgBE #RJ(vy|1%/MS~bV?Spf:ɿSu5M/6uλVcSmiZ׌Fqoƣܫ-o5wv17gtf:׃ı_mVp.y+Z48 }1s<<R֣٦/xՂ17·r2_-d@a8HvXyN_j:1-~yq^ %M%'{:#cw|Gcq@08|4m(}3ogDQ9\eM̭YKB7sUzSq"_jz˛(Ѭ}Pn>Z''b8Ũۥ9o7tqTՏ~;]=M->YQ!/{Y~Ƭ >2hՈ$֨afPWQ[-=g8^~8QeEK{̼;ow:׼σIs{=ia]7{7xT;'oX#j_d8J@2Kc?HV1He)Q$@=y{j/hAnz'癶샠FRjX@ɢoy[mS,ʄѹ$ʵ{a4_|KdEg:cꃳ."¿̟t ,`p,DuV$N-o';go߂^j5eVqՃAGk[Yll෽L#W٧8T"y:{7=b1_*:} n_G7ћy>c;bON#.W3_ {yk-f^Y&цnNZ6:U!k9)Y$8nzC R5_ 6co|cm=;K7znoL醝-] = {BS8dݕI$_Gm;䤻=w9ntN*mUX{2 SKmϔSv5 + 5"Tb@%HOP;"d:.{S J)HPSL™*]J9.c2>܈[y{3Y+Nf,xIi_F.hLSٓU,ɋXry.'" q ru24H)TwSסҤM/AVpVRX]΃O{J/6K%W rIMx32r`kn0X C!W/S/u}(X'PjyqPL]6oaO^(VڲouK?[Oj#Ӿ 殺6w%\+nkBOgO+=>~d㺆9pՀdt b: *ÔqN`3j@*\ /; 9Dr׎)/(Y`V(w*r" E 2ECdW^ IH p`c$L \XA&<|J0I\;$GPd :XX <=gT\G :Gŀ5*Y &)uFT>Z.c`t) W{&X1(^BI.e]*`w$fXRE8|Y~6E# xs0 (hto'g/0=?Y iT+s9GFrP^;1(ךZ;}kEY^qo /y#q2 񘻠 6KsQq ygLXݺv&aL*Iu* &3psI6 VJM qZF( 13:HqPe:C.|Ԗ)X `>*X+ ?V2&gƀ1&q"OE O'O9)蘆|9uъ~ hS_B-nAB6G/ol}j|t&LL%{A(0B؋Fr/eF;4 P)QBh{W_4WWRǴkWʟ~Md 2g!r Rĉ0P `;f-;o[6hvS,@N)bR3 =#D0V(J:2bp^G,ą4^4B劍;)(9WNgXmRd "`G*7IjN:sBWԡ\6kNM|Kz3K_"wWkuMU"x^ZIgV;V+`~5C y$*hIqXsxX#El%7G>z2ά"[(ZZ](qFKa]5-jdцᐍ]D=(j۳B< %{ %jt x&C)N_IKg.'K.&_ɘv1J2w3٩:ܲ(%TĨdE3cG9SP(4l.(%e4,Ѳ msYHU\Hlxqi?K=X,/^DoOPIZr_N2k{$,~vsEOk2e!$⯽q͙G) 4\{X7Gfޙkc&ch1D148=n:_u}ӆ]&3U7fN*jN:jn.lekR_uo~:)1G2)̟oHcLȢ crzyu)6rf fͿbՍw@`N易/[k_-S`d8K6F E^gM}۟+ b8*3 ~vtveG|gV?g|#(,#sH*94JVəad}?~V*#}I߆6mgug罰7s׻j0f=>< ˳:Qq2`}_Y,->.e> y,0qjM^7('dھK1E{žryg Nh_$Um7왦[&fҷ=PE%oolRKY{fɎ>WM #T53I dj&j];:6V=f5]k~2<[ Xv`i7D W37<ݾ42+F{"b &gFr|Z{* NieVs:9dRI[ʨh!++P< zӷlg噌eBq Vh\]Xl~S7iۨ)&)e@!ˌQp:h3lU&sGccHlb xr{*|lE7a~~# jEr`sKKT6bʃR uTxn9Kqա >ެI^8;9?ӁY?ɥ1ɠ\0DY )T.,K)wq T*Օ ˅){u_*~ZHHS,+\c!nDƆr@ަ) <Н+h'{zs8ƽ8)FȞf˱n[M0#*O?k~"B?*\6t`tIn[Yhn7,IE^y%7qX^mvؚ`G'w,tRުe&VK.9lt jbv8DlV|.M|!gGbx1S3]Ľ9Y౎˻Ů.%A!sTrS Ӓ:RZno4 ɈⲅHDf[쐳"৓;n*+D|Sǔb)(l9;;#`[|If%ϭ Q&M8ĹDz8QQ~`M`^Z oCh4}7N ?ߕiNX䢮ywe[X6ȳ%QMD0zv7r<ݍZ}f;QI1?-=wгӳzL<+u+r>*DSWoQ]*g`.٨D碮ZO޺JT6vN]uE Jkz6 UJRv*QQޢRɧv CquӭM[Œtx䛉PrZrH0ۂ1 -)rjG\Jfyˤ4%Lᨋ^ӻB@/QM? }ؠPL⫔Х7c~()?.OgLAb@H btʆ#uj{؁%!k=iB5H[!1Rc0N)1Ԥ˪kC;299pc6\}Մ[m(k~Zs?ʋI`Ip K`P0+kXE(Z'-1Q(Ib~fu>EmսaNѰ)Z9n  YvQg>Xw=ʘJ0`ٔ`$rٔ` S/HT6-] ƛ)`Ls֣xDZtXg&VQ`*$<,mj|N NVF?\W>ᴴߗ%j2N q^&d+IJ3=g1kF'-4yЂO Z0eZbN/{n𰐏qšL[be_N6fZxr}W@P5lUjoli#IJk;akϧmZ-6xZ)^{-_s(\T~BJϝJc5ktڐq /%*S;P+U>|'Q0Œ1S #;$RJ@1O ^ T6D5,C0!B*Caj,k50{Pil5ͭrE\s6rZT>6 %ipP}lm;P鄎&ӷ&Y=wńo.~pbΔ-S׻pGrNۈt,.i}5ID]16 梚Z;esG L*.9>Xzw, G>(uHAW'x nC3F: lp6HF7g);sQ '>cL7/+H25+SK6fw`gRäfL&N6/ DŽu3GN'X0R</񆡀;Kyt)LyHTy.(C%4xa:H&& 5Sgvltwyx"2Յф +Rh*e1lk5?.h9!Xx0.Fqy049sX**9D2I PעUhyڳ5ˏ]>lW| |uz>EUޯ<#yT(?y0気+(M^;p|Hj4fceH1I .y I7 #$KU8m0siER(!‚*8s!8!xt;pߚ.cLRS֓(C7el:UB4jWbC]t4h k&ڋ`B8+=́sCPQz&HЬu+¥&`BL[c*OGӏAD 5Bja,d#ed$t$A`HD ƍQ >%(TV5\*WER؄78>H]gN)* pa2XbR>B6xјI&N)|{UE+dd 0EaAqB$A1#v)# @@sؕ(NNHRpY~ǐ7еuV nu#A]!}FBMGDQ՘zGD%:*Ŏ\HK5hBU*8q٘Ϻ. dN~ Wa (,,r%j ;DY !$tU+7W7<ہ 4~Omw{+Fy~U(1u1%=ù׆K JY֩)y/l)N1A1(ŘRkDe( Dp D8ǭ1&Ej֪tF8v`\oL^,ywӢcn D0b{w{ "5!gl)C* $s$ҘtzJ)}x2)TZiצy8JGB <қVY+AŁ۠jҔF~c:͓ 6DžjVADTPCq\Ys{ނA2i*!cFbL:jÒQ"03+:|RO'#ΛkdpH6v2r{viDfr($)B8S[\Ue XgLX`Bb#*`1s2*:5_רE|7UWs'2⽄y9ZZ g}4\.UeDÌĪۻ1`x=agҗ˾]9 =? SPB+Ů74Tȍnӝ >cDu[?]k],/%C}hÿ>ge{ӎ9,}k*L B Cn]%r޽M ,0L1VXy& D0@o |Ks*Ys'Нk.Vb+$T p ^JT$T`\ ^HZ+~]kIO#4&HZתzD}? ~ʐz[v&o;x<  \L"YQA|aBqTc ft[µRQh *0Q>5|fv5ƇhnW}(ST]ӑ}BJ~#Y;yF/fҫ̴~G[&\^W]]BJݙOGO( jq*,8_܄狇~r`oX 7&va[ GgF|;N`csDP=$՘tsTkr] $d0Ʒzƴ7L*6~kDǛN3 4.o^ V捧MS. rX!XrLM %T V&[Fa4\*<ʔ~PBx{0:m5N.@iI$֔ ߣRNA$dLuJ/Kb~k]]Cxc.dZgWs<.1PZ:]x;.8JSr <ʡyv>pZz +X1dVܒ Y.tjɰ\Ձ?!0Jcv{A۹0qaRD. yvଏ * څL J-Ѣ"EL0&;of7Ӻ>˲:Hocn5YxsP0vY4a#A29EwE%KlAS*;s_7 ~{w;J{} =K[|(#amH0..#KuWH ,pK6kj )ⳁ[=|Iġ5KptOUT,ȟ4iari!\~M__iUze_kmWL'AiHi)g=Zm?6kK60HCda4K,jDE7HHJ{%~_󓯬ʌ TЈ%JD'Qx`?сɀb"PډbsicIc+AbLd2HGDjWY.]w|] 2o/|J΄ң Vs'AHsDQ4R`ciPђ{#F^a#k@ Djw_h ,6ƼꢳFo2@i'jiB& dR AS6x ^:h r*%yY T4Bf!Zx!TvIa%k_vvN~l= tmz Ym`5y@j9/Oˁ"G'!;P0gع:Z^Smb( &X. % T nCلIC0yСl֭]Ϧ6佄"\(!/F|dLҒ# Qnk 4 9KDҀ|@r'OAk-'P*:Y :vrM |ݐ}xð:Igf""Z *G%I1;sX*!:=֛67hB"3@$# -am%Fl`_]5(6wy9.?nUBz1#HGJ!_P^hzT~E"NwQ`1ZqhGpSFX MA86 B-ǮJXrq98ҎO`CA" V蔯#VHk̬ɺ$YrRN"12_@uwDcY`w'igŏX4iɬf7hif&5s6ݩ$}6C16'=~U?I꒬qoF~Fvt^Q3ԫ9|K3hե]IryRƽYY֖g3rnup}>NBv|5jߎif,mdf=3Wguk_,>e:M>5;ݘbݓ$Z~(*^~>;oՅI-g.72KZۻFR3{zF8^MB\wh:, '+>k-%+]> .QUd n7<0_*OI[1YJv5.2oiKN8~mdG- |onrViK $e_Xn}MVWe/V?K6ٺ~fѼgQ-j\w;ԊqyQf˚F{c<ϟ;FuxGLs{:*[-icY޻wvoY+oYC-\,J$YqrvRd(iQT*> iZH8 Qa (*ڻVՂf."PED+&iS!IW%:=K/uQM?~nߔs-V6n0su{=ð0 _~ Wߦ_~ ]/(t!aM7V Sv& -͖lu\rkX^!2MHi, NC[S%eIDlGL]:= S^EYu3HL bOvqG,K6|\~˺Vgvr쭌&W '=Sׇ2LlWmE{OhLmҴ7X;uʏgo+JD9 zɰll' yڀ(SA:eFKXR*!2'JEKԓ#E2UR|NT xJ='$53vqfNtƅYƾ\h;B;p᳷[_5ʾxCsfv||g,.44~<4>]}mP9,AMW!KlK-i"0CD92]Ф mOJhu8h!XH\± W\DDgWpgp1Ik!f_X3ؗcƁųa`72Kfͯ%D-;,S>,ƢgYd$pPB}LgdIYDb<,s}BI"rP.9?;8Ar!co+ؗcFF|2 X$!"rVl1 wJGXfC]@)#:olW5]Q'U*τ"{3ȑ%Db;;n:0/Κs)ٙˋc^t/`,0H)Eas@WR:ZL9I|’1I fžagֱ]|;3Pت?Α[2x2uc-6 XZ0`KhZ0\XZT%+lAP<~cwlkcߎn Sn5nF߭sK+d]0gF =OF?7PVo磼iljuɚQn>u[i,M3` ӡţhQ@iT0kOz5WyUvG3?IVaCLpNFͥucQb7VXk|z1,Dh)ץhP׵.IiYJ_^oGʖ7DUwP_Khd )*~$"5 4Hyo4va0{8jMP6Yև'pM0񠛚IL |kF\hCTZw>edQ #Fб$I#{ gnvvV9SH,vb3`9B ׉c a)9_Mmz{",Z}( 0j7FvCt =+]فz ="bN꣡ WcAJ"p0++(1+,=p_tѓ*eNW%@W4IttDtŀG]1\'*Z}@WH8"buhrW`OWЕqʠ:"bh=۝Т}R遮^!]YgQb$UX誢5tUQ푮^H{T+ :tUQⰑU8`c@Fg`N)|MOn|Z#i xTekFUVQFi@_MC'B;OW;vvAġ6p, wT^ ]U*ZNW%ځ^!]/$UlhEs,tU:w_\߂^$ $ۺ ܱUE{g%wC@W;wC (>O...Gfm[slf4KWw-]=s7jH_xugm9o۟Dvo&/,c|}eowŸ\t͜NjuJӗ*_~ѿ˚+o_UnK,VIp4!EڭWˢ׳[^;:߫jOUۼv_~oߏV"R ژy _'|f?>rEG>w}zf ^c6Â:ZowxNLQEMgdt^8 1)…9`H; EqɃQFlt>Y'N$EI9 ## QIɎtgY y1>BVi20a&Cl2t[T$瑯4 g no?cW:܋SoC't;$YVlm TdR1l`%;%W QRQR"i@5Ё\{6U(@:Y:yW6HS)J6粕^䔅-`'ȯ*g\XbƉ̔+G<& ͒B7)T>åbDs^sP/ EX/N ϔ3:PPWa4c}1,Ec~<)L"!xؽm9ąl?i.Y5Q@HBzom#=`b۴&1O[,<~{I%٦,9aĎ8;g#R$#Z{1`l#kTMmWb*7]8iTލf_ t( |l ~| ߇+Y~6su=wfx36Mksp6ƏM?Ato p2llzpa>ހo-w]A-E5 jR{R RI Tz!&w$p"pI9#47 C!DsNi?$ΉEKď;#2J]PJixk.x±dHIJ9 4Q1G E-GȰ, -J)X|rv;@B+\r?%|X Rɶj޿k}=ζ{+) fjZ]ɫhŹuﲛƅh~2TF3HǠ,59~vD8' jDXrIaΚ9rD)/dR 4AҡV'aR/1,DEM-a$EV 1gkxbǓѻ }>E|0lM;2߭M1{jt{+BcCwAkDZ +" A+PRDի1VYE8`(3*h2:ڀmT Kb08.NPL2mGg?0c2McSڳzb=6EŮ,X iAV?>3lQyl$ŷoLHՅt!";CD]1%E,+i is Gs \JjLd+laDmxb0`ȯۈr0v QEln%9Y;1iԃ OYCR))X8<40˜)=f4F3.,s*ejEمr=:k] }i(s\O_MHX̵Q*g er+wrÉHgISa4O" DDz5Pn9(tB(Qbb he-'-uﴲEC|RNtH#AʱCGGBGJ*~0_sKj,; 0$9؋`6`Q[C`p&Q9cK|R#)Ƚw,>V9l==hc (,,r%j ;DY !$t/)ac0ln{_ak3MT>FwOŐ SB ν6\ BaZ) :44"伓k2C' bLZ52lE,`8q"VI "5 kUkU0`"{&CDX{ rsR"!w<-,xG AbkÅf$f 9<[rWzWw׻Yu0Jm0b4ۛWW[jx<û<}(>sCSbM`xALBz dU>C8 Jgx6;k!_X"Z:~k! </aRu*n,$%`ZS*$&~փVK9Q!cdzD.^EƏ~}]Ɗ2_Tٜt;.@y\;q˪g\}PJ ؟d\sbl5EyE]_łsPXQGbhΈ'j-#`&tݾs-{Hfx=*+ekb03ؗ?Xd¶rS+Ꞁzƚ9I.s]ڠc e & ¼&BP5; fUTsc%b ͫz5`2عT=\[3*lDV#P)ـ9dI%Ne'ɩ4}ZmͶyM5h46Zv 2{:cuNJ`i)'rn7 +¥&LI8FҎY | \KPoP=^#Z=m4=EۨzVhal /gvw' eeF3rt ! $t ^:cG~_aZ]@͗h>5-ATxGDJIft0KEN4N+aU'L>:ݗK½ P9t3^1,'M]7z(3RB$<"Qz'*")pu8ҎBdQӨ)IX: JIb5@J"*v*KV(sQc{2C9ro+ Io,*d8VB{ S((bϹqi7/``gMMFYg'wM8a<]/5LNnb-l>`KNޏQ+]~f/PT6ЏG}Vձ\_gf賛l6Ԁ)Dty5ϙK߽sNZ1`Sdӡ'8N<r U6:LP_0wpOoV2^~ecYHpp;܂5?_7?,<xXΰ]&d<XT%Iϕ `!e4)I;O+[ndG¸7EbL(~yp`[-iD6w'HIUE Uu$NbIRHc3h MJ G_JL`ۡT"ಧ f5ow)jY6g'){0rݰe7O;m01RbmOx5-&EbHlUZ*pfz&~gm["9@iէDONFBڴ0jopoM&nE76P{QX-e2`f=\ln1 G>zEioqZ>P#_*iODψ[Ta'% t~ـ^j5F.H'B=%qb$U9̵@%# >mBR}3copgJo\؛eʅ3ą~:J݇ {./._<2Ow텿LЅO󻟜)J4] %DaT½t٨0d.X(\aSCsBʈ'ltMFTŷS쫶-@-TJ=3vopgx}7\̡voq(kSϬMkO I"bZtATJZr` #eN*xR1, jC "XEm{WM!\TB]X-b.$Nu|pOD6b #e2"̈<1ĈOYǯߦ$mi?dQы} )܅xY2n>o V}aܼ݆n~w}+GNS_lD۶ ^,-^nB/faS^T_{M~xwu6[5uJ;i{,|C˶}ESwlXk ݌r@e$DU<|`˖V8L:8js9g*up^2ɑWV;KoA.]=Snmϣ/A[1q-lrpuޢQ@4 ]E rڛ/nI5b}K~v[p׀/Z}$RQ׷U.ڀHJy-\# nK*m (g,vLd١ި91D \>HBHƨQ1딍FfTƐ}dR4 0`R*A(/u=L$Ʋ84]Сْ9X) J>R)P͊#/%K-Tk hFT M."E[JѲ} 9]J6[չ6H)/f]]L% Y PK2b!l94fƪ5XHz)5GQtM)2ahcK}O-YLbG $yG WP[FIgT0K@SVLȵ2+K(yFTV\j3f0ypUJ7x,@l'@(.ޥAڸtk􃔵M Imh=ɻ>*IXMBZecI"dh+v@c$" {VҞ/&a2$eh aS[!aHPHa[*.R,(Z-cEX:IEzJaI(J"Eo)k)JߢD)R5kQQ"Bh-i,`h;4e`Wz5%j%;Qyb#V2)h U+11mi[Ugv8ƼMڪ@p0ȜaH$8k,="9ȿ$EѱXVbG;bUU^WWt%).tlhhXWoT\s ^-FU5\+GR7fa=gA&P̡  "ٻVLU:I=RqA11vx X-V-tC*D4`Ԭ4 3*0 eh8&X( ꛐIv4~+*ӉPu5%M%$`aW+dWVb@,AnTzCZq2Fͷ6Q1j ȄكHc7uӝ1(QdSG*nYy$Q7C7a?%ӌ/%RC%:&|B(87ud&t/Q*ACmj((Sѝ E4GRF; U BQ{OFND)J6No& JMu2R".5UuA"F_R]"k`3/z $$dAђ@{ er1! aUh#ǻ=sA P鼍/7vhH+fD$' GcE( / UaZ@˚Cz}NO׵^tjV mCka&a-A7 /ፍPtdUt4CҕjIJ2Ba1%OpHv9k=/3*XqAyjN$\Pd"U̫ fMʴL5c4.zOC }Y%H Vx$ۑ ox.JrA#+PZDbYьmC5YK1Z1ة`U4tC\,[ͻgAK Cj)M.1ڹ^y6gK. Hasvuh{uT$K  n=CnNN44z63 Z';-,F ݚk)DzI(ΓDCkBm1&P拄_ ArT{R@k80)QC^":$5\Qr#bs"Q'%\IbLvP%(H Ш3) @)w!kQƮXBď.w`E^QH"NdviP'7 B`~?"oQ^1bp0) ,*1"@jt01ـc[sOҕȪTqQc@sjo6i]0s0Ǖ]FzPIQ{/AFL(X*-lF]Ρ~]뜼<޳NKqנB]h5х7[A[FsE D+ (|u* (AFW6m4EO(Iu!~)AH 8QZ{*(=uGPrm⊑4 t h zul) lKC$`PrAvƒf5$=CH28DԅKq$@Br6B!(L ΃rt `j?n۠˛[qr 7,c dc *ATPܛz|&emvH\\@ps~NU-he;Fb˫9ժ]/ Tv_= )]a-+8<ޔ1#Ͼ*}e!ݝ%-t/-X._ζp^Fo>t0ǘϾY8ҋj~v&.#o ˫RHpoy{W+Khoz&-M[/ֳͭiJI=}2S`]9%G5Nح֜7'=Iweۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlzu9Vt p9+Onڲ)ڭBndۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊VlzuʘI٭{?Up2['onVElbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vn唲>NnECl݊9zNnE(`S[fۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊Vlbۭnv+[݊VlnJO ^4{KxڡaZx VW[sìP=ich{J0f\k; hvs K)'CWWuPtJݔ P++T~p"N2]=A^ o&DW, ]\;uEh? f +kT_"LWOlR ӡ+UM4 ҕ )v) n̕ArbIytE(`zt=eBt34]\BWS+B `Sߑ3J͕x *jun{;8nhj}4Tl}J5/̽K_`ϫn00̖n ׮z~XeuvZyk<{Hc\\AvK6`1Kfͧnثj?05Ԅz17}e64Q?z gJ5ߣ]]NKR nPVg^~vNkJywe}g.y0)!ڌ&Z1!#Ł}~WGfaC8d=Ũ[x(Ky28Q\A{_҃QU Cjt|pX,9cuޚm6Sz --YZ\&@h:B#fJpӱt8bbBGPzס+CSNWc[W?G~h#y$2GB݃1!`od:5"!:]e ҕ]҄ q+TFutis:HWFЅ G=b' ]Y)C++}lytԞtYI唊ACW'Nhɫ+Bi=+3rBt^Lg~V܃КptE(=]=E 3hh9uEpdPz.$]Y֮NρfR̍ \{Bu~xA7Az-"vMms u1٭o.~lV:ˋl﫶N8{wmdpa0-!k'E[oNHE{PDP4'fY7[ `w翵z#{ w~ -/ZZγݼ,_ܚ{W߻#>DD]2Jn}onvv咄nB=lhF{\NS 1]N%7MtJƷ.d4AEZYoV&giMELE 栵 ƞe?e{op"Gz|o6 Fۆ3Rqml/ n^}?w[oW`mhI㜵7VF1P/AScIg||sAY㆜iWvU+w3&JO9_z}>SPJ;3%NHrɫtĴ[t}.bv{OR#fUu1"m"{ת'(_,/ srwך4aHdlDLۢA?>C$/PN^Ŷ- \:ToIIq:}#(ׇϾY8Ћ &.#}K)Vtbux4Dy{XgZ/^4:Z3nm]pRR•:g[J΍KzBP9akj1e(Ă ԗ(b%/C=e?pӎ{F!mpŀz]FXӃTG8[hnhtanYr~j #VN>t\ȏwpHXX_S4,sVjCn9;RvT*ӚWEMj)TӣV?{Ǝd>mE bw1|X`g0$L4%-}:Vp*S*搳5EsAE>1Zg7ɻWh]1m6,7Mb?giNwX`?϶ xy@+Vv;κ1_TTArY%8 `bRT ڣs֗p^ XրS6g'A&HYGV'Ql4+jWҨYtxd+ǺO)(i$Jygxw`辳zΙf˩6s|yy>3%մ>z}WUY&=9or}HOqco|V>jͿEI+&"h$~A N*RSh4-"MONTH%LX.Tkl8-c;6smml ` _nGTP]~ni9Y]Lfb[lI9` kg *4SI=4٢l0*1CۭUh~ B1VMe]:ZMNzr@_ lI䰱n6;ΦC1/j7skl`{P0ښ` PYgɚ✂3W_ε"`xH4fpC(bLE,-&c CTID0f>:hj=ic DENRLP&Zp*cJ̤v*8Щ"6 w2vqѭ|ͤ\E?.xFG R1IwL9 gu ,ESS$B4]| x)teݍ-ᶿg0aۨZ|^,mF~wkk%:{}nRqRޘUE}BJƠ3ͣ7Q(]?oS 1좏-;$;'2A/%9!$!mߙHNfo9xvy լT*ɒg`%g` 0TtVA (ER7ߑuKWhFwa92,%lαH .Ic`1[e0?< {{>ł6O*ǥ',fɠBr)C MJ3R:A(zT&{ڡ5=kxo4Hm dYg@'e X.ZL%䰎4Vs(RTmo+(uG^ƿ:>wvӬ/{D*?'rH'ߌ>LDZz_hvg2[ѿ-6h1) ?o_[d¤jou#\obcxI͈a2Z77?~rCM'3/Z|a(<[ ӽv(!ꭳJJWaȷqje*O|A(GIC2%XZ>.}9NEj't 7ۗY V_ߤJ,Uoq͎Ox0ge!`aݻ5 8gp/}*/;p_zџwluxyGkOsmg}>V ,/ ޴/ 6%(sdPZF4A*H/wT6IER 4o͂6矗Ȅ.<Ӗ96Z)dLKOdQ3)d1FZҪlnl^DV43#v7ZD)}~ .,Xt ,x4`u ^' sGlIޔ^Տ$5/wo}ۢRɺvju$|M]U CurDQ_2>ݏ@+ZnF+XoMrh/P,ݾN=IjB1ʌ㬭t-U)NFEFyf^\_=3 їb}iaByT|<-]]"yMp5=vuU}{w~Ա?X+qO/w@R$x֯|a,صYw2!1{}Z}U]S{Yw UbƭH ĕ)~l}OX vLO 0\;J|6t!YJTyp?e3>yp3 52uHD%PZ{Ci@k| dBf tT{~ziOeHUۥ{}-f}yW>YE sFeq9[=XcF$ON7)P1KWlQ6vޓd+ǖ!B>Ep8"^O/§`]U@2 uoK=X+k2(1܈TReJkA2D|R >PLmz#E&J2Nsܧ˗O<3||S\>̭V;{} h~kVm#YC1~%z oWg;βa|sړפ@ZX@gNۤsdBKGz/d ?>/VYOjE5@6zFAXKnZʙBesVDc<={IjRzhSt. 0"畖 lv*6!f9qs.f>_ e-gPߏkR5!wG@bw1߰11γ^:X@ 6,2FoRVXC_BA"ѵNe )(=0pJV$% <EyWIP"$X oQ:t~eڱ)xߩZ!ӏ{v~)Zub{)CSx]_zkVuyW] jte N+LZcڿG_.ϝ5.j!QPQO&p& 2j%0?y߭5 ,^NEZxl{E.V"ڑFkRAG%56L4aYklT) 9+ڈ(V]LZuBգPVGfsP,Z(kوt ey eaF\`+(eW lrPxQP Y#S(*hlVzJ u:]1=bifri~Y_8C;sF8,=7rЩPRBtO4l! fΞ9 0*U$`Z|EY'!Ũ1+r*ѐ "$A$ᔓ&Q{H (jC5BSk7@cl8fx>ӥ2)b;ԪqmEbjs4EGtrdTX W؁4tf r9"l9ea 6!0+?R=I*/2\!|u'*hMY#Hk^ܣdjHL%B2SBA o9ʳR?1`(=W5G0$ /HgݱzyPJNrmnxUQA.H.܇Lv -3hp;P**8#nZajwIigFu]YvZ 3K2V:!(4|~|E3"@:z᳀ER@Rt*he_Y4{[) I+Y? ʴ/3EŪT%;zՂR{ SFf#-t KgIIA,vLCjVEb6EP=qRƠdbQ Q)4@goѹF2I2!)2bLx_4R#ԵIR$1ζm#_Wu eX7qrW[l]qJ5,i'RTuC"E8#d${~H =/4&: tA藐~U6ws^6{m_.>彷ϐt]NM#21l%?2 , WNEx2{eT)hNsB <,|#dk煨 ) qWB >,[AXA⹬ Vx6+ t:8]:άLZ[yp.+jm/Ib9{l }l|afTfwe)Lc}"чcNc1Bl״cƐ/XG`X$t%s8P Ti:zWn||tq߃b*ab;KU[&n񪶦R%yͪPq]ߌ#v='xsЊn{Eg|U)/ q3RfcNXc<"RTV;rD8 /qC `kk \PglDm`8^VLFR (3e͝{ᗹh/Gg>L R'߰I?-//{ͺEnj>3xpcU5x&Aqi&a b*‰ hvD2Fx旫ѳ=Yɞ1 }&GuPv/^\Wr>oZY xy DIާbC(%g3q&X>_G=G9k~]_$uTwWr6^ƿ|Fes$vs '*/kY^[W+ˆr l5G_]"K|G -o}ĖM ScF7)hq \a\`xsӿ`PO?~ZNo?]8Miz5\GOFK,tȝrw}z&T>| @om:?`7B4w:2;vvyW~(u  1szkM,ۢ+O _J]ŏ;PS]G4:GێHGgAݢo͏aÒWjydM[P.mpl979ޕtykP#KC'K[nǻmGNDŽlX^wlN.n_؜sw5a"!3@OZ߹''X~)/a kJR %V*f\+ou*0/$B*/3!,Ⱦȣz!xLk2^?qgHk̈J[ LITip9W2 >rYd=eXVŁw2olL=o3<㢩aY;{ gU9JD;)j oflN?KGX 5\5s,* ǭ5&GsdJ{܈*md)wr؊Ӣay9sLn?}9{8X=Q]W"L˔⡶3YRpV:I*HK>NL?LĻs=1:0ϟcv#sCmvaN0RU гW #WBۚs_gUa] ru(:{3ӤrYlntO=Z/5'Rճkf4M޺hEu ]MCx_s|j6Zq OP53g&7ޟmva6<2a=ppM=O@|u=ݰ~fKD D˄ `U2QL1eL^UJ 1PMW 1 P:4&jBSf&+ku*q*eqMW XS PHƺBĎ+Tilq$6,g Pmfjpu * P TJJ2+#j+v +ZJh z}J'#"ǒ ?o^mt9ή4 p78wU#F13I%2{_Z0 d`dJPkiV%"cz60}L.GK7 ~6Jt3Wp7ńϣ8 ׾O٦K=+Bԥx>0֚;m^BÕU!JQbu]׵|P5~[P&h۬D1kN.(񛡽ZҪGzz^qJW(Tpj%޺B2[WCĕ(@ dQm 4yj2RSbc- vY*-0nS*AH( dPұcUj1=DLe[svrT7]S榽ۅ H-)ʊ\(oK[z*87ECjMZNӤvp͍^J߽jݻWO3/Vydu1z}?\᧺~-u JM}i 7/*/!fqE`j% g/uaU cJS:O^}d4|wISɓZGZ'D$׿B _ ؇Eă?ʒ#q{%+]wX6`K%yԩfq1n-):L(ludȡOTR/NW{ya{!aJ`>AWIC`[]Sz1tpuZ ]1ZSBWGHW&9%+vv9zZ ]bp?V&+fAtŀY ]ݛ,bt Q UPK*[]1\b/u;P&Y:Fwus K+FEUP>%]1ผv7A ?|bZ1U `oS 2\rK+F͡ \{PILZ)b҉2Oz<1i$٣sh^ wP:;ѷI$g,̃}B;n˃9ks8/Ok82K9hwl29Y&ir/hiGRNI =WJ/~6~~apz$DK/s=QtLtV)Cr.K=:]1;lCЕ1V%+uh):]1J:CW 7MttVK1xtE勺 ba9tp#-6(st(iBW߇5jQKr֮uK+F~Q$tut׋RWlT\ ]1]1Ze4BWGHWAkgvŀ_vDj1ڃ_b[r/\t7[?oy{_]ꯧggb0ї:ϛ՛͛7A?l';\iojwox3#UwIg.ɟE͛\m#7/Qw-vWKu&-ĐGn>}#>8Zlq\bs5,_M۷T'^w=n.[̓^Zi\y5]|#޼N?["yqAy@}"j[w't,<<-j'ͷ煻淙'@0_wh]åȺ*-QঠJΚT|q:+g+E7wI[ ɔ\HnUK%htZn\ȥ4S5¦i|OSТNowBC(B%I:,e*:6gBIV&m=5Quali%tmЌv4GIhSo ׋w>CQMVa8 pJ@uM+5RœDhY{n>⫲k5sƖjhjRU=QKʇ! Db$ۃB6ߞjE}∽}Ai!:3aP/yrM/ĥj*٠db̠JUYa:iB0`9;vәx`]>\m$u~7IįkU0o3?8;5͈ G 1=*4*KPtdUm jmlAhqU +0똆'z$;f .HµdHiD(2!Y!XPti,Ѽ<=K`r@!YF[[x+$@f6 *YȩՏoTG}^ yT'ێjJz%B4D4XMϝn~W."O֮ TG<)} ]{ #Aˈ|uC.MAy1k"! 5y]%@_!80f;PAa5Kq[S-Ah[Ry !zbA jhjw ,hͪ]Fp2zv huj< 3`EV_pXxwbnuQKhx?_{y{qծI>͘Y~9|zu3~m6dWڄW Ih~/wZ];mή~y:>=cdSCۭC"Vԟf1۶oV=ɯ=;EDŵ?)OH3<Ь/~s@F޶(˝$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@B|@pl> _h({>$> H|@$> H|@$> H|@$> H|@$> H|@$> (vaI> ^2-:FPb$> H|@$> H|@$> H|@$> H|@$> H|@$> (* v1\ob}@@ o$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@:yOՔ[\nNhw]vݯŇ3, ڇ8~l%ީc[ZJo[b[:җCUz'%:.u-<ס/Y'&Lk!3ReWgӟ~bJ!׸kJNȾ޺dҷ~rN9+M3࿱wm#K;~ȇ$bI0I;9E(?r&Ilږb~cfuw.Ezo.i5RIJxáLJ'?1\m'O W[I˞ ݂+\:0NUX$DWIڇWgWHUX𽁫$TWIZ vp *Eŋ}bWbU}$-WI]bQQ]%]%q%J*pԝ9bGpbo*w2vtpx1ЁPtU1LJ@%__w>)ƇP0R֋R:p2]AyDHؑN,C3F<ɴHTk>#g7*7*|lB%lf"z/{7yE U`zv>U>Z-͑`hݑ+s&ͽ7.&orptw2aka<@nNPm|3TiZ*miIUҀVmhoUM@ag$)}BTki#29Ᵹ/"bw~ *: ]|MsݰL6 *|L~jjI\7aIZ\-I);sj2r@`Gp{WIZpRnY•bq3@`'4?pb wp Fx!֎}ʄ33H ;DVJ`S-ڭ\<䃪R^iW3C0|!Jl01/AРZW2-c1OT|:OÔ||$ws; 4 >v\yM7] Q~tlg:ehvi9wr@uʹ)³mKW|X>SnRU~yY,f!YnuI:+(K֍Fyx F 'Yn{X;d/9v 1|%gStJ-sAw|F6(uDfhicaaLޝKTYɵk?%qBV|sRu}?X0-&CQOcYWI+{\8`'lW=}+ 7p%h_*IA<$宅tp(p #9{WI\7p}(TgWTEUX$W F|*IuWgA$è#II-f%\880#x4*|yts#/]5)]̨6$8hL46":.)箊HY_qq'6ͮ6;^&+H0|Di̥4%H )p)aKS*0C"e6D }O^& UnA4V⻪h jVY?:Rm{0J3BEAj4f+j*LpP΄f?]0 }bqRqK^)]|8ޣb tTְ3auMU)M*˖W@T{r+~ůRZXԵt-ggܜ$z:汜ZLog>/XVmQ3N{ןNp9ϯzl> OpqxY2u=[B嬑~oKHW5%ܘܚ0 ͑̓X ƹi Vl33(_ʛ6 u_?oN܌jL-F+"TQTQ-J@Ҁ:)+@<'3Lb񆡀 #LOni'a)+"QGeʻ $ )NDp'KSByYhur 8M(rn=ZGEd(rbrX+卦ĹY]Crq~w\aUԴ>Q]L=zZפ[Q'X!̸ܾb:dqqtzsߣR@? 3 V$Tc,59ώ~vsG8' jD XrI25sR^ J? #z!U{xd!R&E.jʈhA #(H8HL?[gý5. h4/!mv m񻬇w|f''uf{lv{ovUno 1i'{FPH*J+%KTQh5Zk 62( FXR*:dڎmK-x)%;c\Ǹ(&qUWz\z\VkVڎ-)o^ȸ%ǜ84i#J6Je ٠LfL|G1 YXi߆ݻep"7PgX+9OHP"A{"hnxbqn^7ihm#v?Uf;hcפ"oWŚ$gcy[cFx\EN;z"vę\<%vT\ tjYdHrH,3[ Chp&S)g,wA.v0ci3aiy1\ B(r,WiA& ͰHQMQ XIB դKXލOc|pZ}M/NjbJ[W4WҢCϑ^=.ESC:SJp``nj JY֩K;DxDHܨq 0<0(ŘRkDe( Lp D8ǭ1&Ej֪*0`"{&CDtį[\C_7.}9 Q2cڎmM \h#9P!A&DXxANB*GbRb.LN.X.t06ܻG`y3F~Q\P2͘'n$jJ9B 0?/=WU;>?}ŋbheaRdcugJm@ "|NA|dś^g)g_;WWi俦`MbU)A$)Wb$uN$ޟ_ތMJ_ 6MZUvk#Or^vHJi ߟEiL2dRr1dfG)T7U`Fg**"oLϯW|gUQu9,/{>L=P[ªP)sʼnogkK)څ̚=9YN"F׼)4u:GV*ۈUѕ?lX)ORҾA71M57K<̆rZ7|r20,Sj).<v1wD͒6v(6,ozkT~j >H09d6N^AM{ʠxPjʪDk^4u 4/Zwu,8 mxH:]=@MMgN`G'M=GGL+BДcj\-RXE G ­$yFamvE/B<0&EZcaӁ4մ$֔ ߣ|jXRN &d$BSHR]- .@d^p@6Ƣ_eƔ%)AUϐ%)jJ=_dΰgꪧ{W o!j2g{̗TS뜮-QWUą'o0C#G`>W\RŵF\ݎfQl1g 2wy 5:i=Jdd"j%'5YZrzfӣ\ fդ|}qfU<*{_sy~OGʋcZˣ,W[nYgV*?_U&>!Ѣ&'h2f+P'QZaP)*Arْqq>q :~ Iflls&@ۻ0f̃ 3߂<̟51F1$7f p} ڿ߿kNy{nmG1YF G7l~ 5M >/ϛ>lPy4vE;3$ͬ\eS CYMg۽>h!+}P־?'Vb & ۷-Xtߺ|,-`<{IPk Z|}~} ~1?rxE-avr=]`26nPz) ǝʗ gd9Uwd'wG~foB _)oYXsS. L|:@onW7D?$z=yhzF.c:fK}^ Hشq#z.iZ}8(@oa:0XρX,n_01 J s@?&~N΢=  eYնnwV+Ov<{US!|z*l^{JG`Xi=?FmeL69NsF|܀V*dR z-vT0\r8Υ>0T: x0QOk2SvcR2B'+2xAàeU`hR" U#)E\֬[/\]Jg ]VΞ9;M@ťbBi|'N>W#\yZsыbN`u[3V0פنϊbr"!Dһ$zc-jMiEM,%nRcO~|3fJUo'J29kw|n-.D{KUޣq;0or0[fXBE%K:ȁ G X2j49=Ch%dzR@*/KsgR\KG .mMiXT$P[+B£,Mfo^$ :Na޾uNhhth8}[BiU0ɯB6к*i,Dc,>2=($OXVQj4x M&ߎJU!e ލ5-+.Xjq6v`wJ<[Z9=cOL:I,b*#yp>fY :tS O)d,b-8"beE-.BE py2KIh*RJT 7@0jU-EΘ6@Wlw5 /рiF& mh4(!Aں=?1 e w)q!X`H;tTI$5h,ѲXVΞYKkzB+ +e&ВBO~zs; i˦ho%֡i*x Yķ--:!jHA:͂iD qM0"ǒ`1CtuVI:GK1]|Efi*,bYzzYIC 9098hAsCAfcR ܗc+P$:uD*٠uܹ`/ѣrn $a=}5B$퐁gً]e})27AYFGBnFxK,&5s`x!PבʊtH`,d3T]yc׺GF^S ed.l@ jBo>V />W״q Kߕ/K!l2}v1a'M? uPt<#:3R?rww}y6wA^XMr%V .w㜅/]cӭ)U*"v;y|ƽV~IY#(]o؅uVz}^٠mo;&Dn!H h0 OC;E!%.j?:~a(xQƇ!|x,$%7-#GrJ؍c\2q6fgFKe8si錖4_,9 ۑƒw(abN7Ӄs̹|1kyk=_oUd/.vJl4-Ys]-& <6 I$pk*W.;v%Z\ڬvD=/rcBS5Ih>$Q,ŋ(]ӫv|yyi,}]eEg ̴Ο] ?>աS#vHֶn]708xhVqHf'PAx 4Vhk$,FIe=x{i/hnV{>'t╶䃀QrKHV e6̌d%-`s`It&fn4px !O,tmM,xt%`N>2RI6wϿYL746@RFs& =s(%(%)%(%}p lك ,[!QjwE.5JX!9kMeJ\ڍW4zMKƹ n)pKpNŧȅ0RI3^C$Q[WȘ'м땷נp>66apM9bPO|7ЧZj4acN5.jV:lS 0j`5m5ֿBѫ7&FZ'BwVbJ%}[d>0sԒ覶B4&п?BbR-W&px`:_$"fpGaV>et>Z/Nά%ޥ?1v৮Q|x:,h7!L[ s;O'a 3Fvwˉ#,,3EB"K'FB2 ]![f>'rh4}K| bO~&X_]-pl6y?9w >Y)cʎj߰]M.fNQB+s^9knPKg6c*as\8󷳵κv:_ξ\[,C#xΩMoHڬfJheOH{=?! WtvCg_nwOzĠs(6OlazGy01͇{Gb^SO ܭC_zt;2UũfnK+QO|ہJ-uM){i[xCR!9'{A/s因؇dWz !re A2#S/ic|:Tx)rIQ&[ %B &j8KVLTa: eps4hȘ3Y꒛*kgWΗ*HO}b%tim>JGbܕ X9dDt佲CY#.}I/i:YΣ7XD]1'C&Gw(~ RM/~sk)T/TU9H#1(XN' sUqP(I!FIeɤ#K%m1rvN^k(㗢P&Z4|%QJT`{D&$!,D 2EÈl!r< DI-!P,VLxaJ;E`!U.a:=CڍY "}$V seǜF :G@4*Y !ǴPe ~I*,WM"Π~JrAKVi/Cw),Dh=J ke%eȃpyVQGm]w̯slJ,0tUy9X!$shKM k"+9P%xfViŭh F8ᄗb?  "lAN!t0Rg-hM'mZE1A x!2kC:go `鮹wt{ Rkݭ[i ^ Rqfu*Iqlp*xk%IZRRYݹ2`9([vN#[rhӴ~lkAP@OdZ&Xɓ)TJ!c/mɴS2 |pFA}uxKmRhRqɑXKΉ(]Զt EDZ9F)_h! +J*oYJ>Zd&0l`l9_FWP6h ֵ ^̚A.7A@>jw,p^ɂ۹9\#J7r57tRBx$/ ,ޖZ{#A6[<\ ѱ3R}/Pנy\L,J &gtB'F3L0b8h0L;S,7cix&[g&Gqc7:[7ۼڡrѿ1EǤ(sa`QDǜ1{Du4BE(D2/OٸU bxŠ5)cpj,OE>2pZ4pʴO`霗רb'>b"53zKp2\ 4S뛎{h=\h.Ve[~"y[T{<݆YwdiL^-7P ӿ/sQ7+2gcB$H1 .A h( UZZӐסU=UOyޚ$ @0OQTvN9G#hGN @h'9`fq*oH3(ZZƤ(Qº -6 .n780#l/5ijO}=^(7I\|Ucθtrɥ$+QTUW2_=-WK^Um9F/ntuy6d>3tT JojN|Y:?M\y9U uD n[;XO2k ;ĝɜ1/L)H鈢Wn0-r8$?_bWT7a=d^{8~w);~WX>tev.Jꌟq:\3xpp.ǿߍ!~q;(.wf!!@i%ct>29eliq)z7ʺdÝ nsΆ?pÇk~/F71|=@3?$S898HTS|ag =UƧIe7,EÌ!e|3j^!N*J ]rETie % nփz%UgÜ07s<uxcv5(gb|Ũɘ{+P[qSP %. Kk!VCfK r/*wjf:ۣ:Mѕ'cFa0)1#r0M':0,>޼ ד4S4AԄ] BԤ-#ώՎ9Ig i&h:$PnrMGfyBގpLOx3潡)޳hްѼӬfahCzlD}٧p{ +RJD(4C0|/p%8c!BmlqWGyd^QCo |U33lΥҭ7 JՌb Mk(3sr5_p _0BuQz}*8>Sm42o7\dOR3Jr 4C$)4JQ#e sg@D"F azci3̵-hy WT tvy8׷X6EϗvO=;9Zfw̧+C*n+ vI,zFʫhd= p8X#®Jcx?(q+<y& )R: +5C!^~N?tֽbX^Ym NJG>ab="~UZ3D4&hUؿ,ZnitZ1-U^Fb ++չ]rR| =RQ)O$$|[(l~cI'ѷ&P:(>TٍWBW6VfJ0p-u8YΎ/λBs 8,Nu2`h.qXc`Ccj#?,&CeDKp [(Ĝ4˂Q'@$ 8QMa9ltNLAEA]^T!%(McL;n(׆=tFf/aPΛH9nj5I] VR6EdZh8լ0lŧ_LPHT5 k_k j3qn7fVoư4*fPs'&:=}Ov$N,P7_ͤ=y7Sʩڟ'MT{GR+߾Q}?QbN3lI;2<1 T8[J@%+B&W ɀQCa'36v44㾾4f{'O_^_Tdcsq_v=oiΏb@߸V蔣 FW. UN;JKS" eR%fchW*ȏ{P +Yݦ26JOpl'/+))k챛sǎS1ⵛi}mG{3aht`K(3rbGX 1xbÀâ 4 GȌ k1ʖ} e̒ak1&8s~?l&W1#ҏzD##&7R"2!'bLE,):i9R(x0f>ZhQ3f,RL%*8hkwjL2ĹC6K{"eА:i}oGx#mV.Zb@N: Oh%PQ5X'D=}šfq/hhjvKEw?GމkϸMǧ~^6X6l>P5JusΨdIӿW^-`^5((*YӃR]l~Y0qz0i߯]>y>g/KV29[AGp_O$@opT0лD>v.n'/iE(6}ê'?,>ݢ?#?vuՋZ,0bv:_8C?~ےbA{'\JWH2 kbQ_MPϯyWr2qxv@#kg;aRJ9FAs-pr%ˋƒ4P93@v'HZ+:8,kRTq67\%::qƃ1$$T_dnrX "h6H F%ѩut!(*! S!| aLt(6]"G%]ik]L5xwbQd FGO2RPkRǔ2( e ;Q`4}ˌ[hAQIV<E'Bqh2ZPT0 L6FaP5[!xW5lv>6dh,t7eз˾ayfGx}4%w>=+~eh9Ծ{C̛ %c2TV$|uEQ]zȪ8WMR*<t[zݟC6hwM1fAV8TClӥƢȡ`YUk.@pz-{m\1^>I M>ͧsW'<'TB}>z2vmP}ꤿ/;~8݁92Kj]WZ_qwJWSunxuM=Uz+.TcdȺ ?B'z*0Cc ޭMS>*]HShj\;bɌuO&| //_e-m(J$LV+I%dAF+:G__W.L5#;gd׸P/Z䧋>9G`3=;♼<`>KYc_EٛY* IRs1NcEEӸz2}%烤>vf>bcOK7N猐J]}^3/{Ke>, ކEZXuzLUz7[qߗ/ae!nHǯ1I3$Q, ~_o]|/Mkǭ?\ yh)[Bٝz|zwRQ嵧wxtǣOϏt~Icniwo~ yd#WEw?M3_;yc0No8K2|{іy.O~ˉUR-Dm k\j]Svןn{c~ӏχ+׊'Z Toև ^tп:<`Y8Br+u0$8ꛈ:ͰTq/Y/"Vu*8@^E2 5R֕##jZVkD*e8aݗ =\E t \@JS0D-6 k)2UZ30E_R{\iqfJ5V\l=6L5gaj jnf bFDxU+09^^\7F["tg4w=YcxN0%4^VA!sd%p{;?oi^/ ʁgjP @TV-kTG""Aж4YC){.YĢzNeV 2"|c L;bHl`-[ٻxy;_ۓյy.|}#jM`YlV[VPr5m% Pwʱoku|1}0f휃^}\++]DwFh} ok岴grf|Ua|2}z󃫹mUv ڃ!e^78ON_~?_FuuŽ{;.M_y#`Ȕ.HJDF٨ L6#XI$B&?DWË}ɧZP=y#~lJ*F΃*b˨.*@#?v3.RqǾ~)USI$\ Y%WAxD2>Q2!5 2US@IXee42z8C"rvNZXiph$k ,4޿j&W}Q*)pAW:TJ"O&٘-IdS:l\dJRL!dX#@`h)lCpl?fj<(:rtKaP` YC_i¤,f 𜗀a|ѱ+`.tPTa4,)mWC6ƠGhR v ul^8ų,krzǴSHaV܉UsnX &QB퍍›S~Gi~DA`6h1aw>;;.- ]rxcWd*A C+ z,t6ȸO'L?_ 73sI`$E=HQT!R!8o?x}tms>zcݤy*!SA*gsb3j^UdUi =;J0ƻ \\um}z7oȥe@eu'%YfFA"]&[]PHe2d1Z+9R&`@ſ-s}q5HtHʇLh:Sj?kId*B}fܾCz@ܺ~켫Sc*}|~U~WVKm1FOk)oϤu[b mڃ+/~67 '*hCpڎB#? ?}ZJ02z!fg尮r`YScH9Ң2ZȣMh= |V_v[uOhC.\]O{t3~?V饳1$/vJڨI9im0RTٻg8Rm]^&@ ^/1C#I+ d~0n'X'A3E2ҶC$͇Cǰ%3iT}Uu=F,IR^EJئ~ ȳImfcVZqWϐ6Ŵv,Ȳ`Y_CIc0DV'b&R+9wilHehSs_FFAZ|5ڈ6moDnk֨=ŶeZOĴRi-8KXUޗ8*y/.QrĊi H+u$ :㦥XXK{9ݧoּs@)@ "#PE,G"_Lj%Ap!q$eHl: }_i/+-z![|v!FSr hӴylQ>oB#g&#AY9OZQm$\=bAY4CV0Uޢ8|i fH]g¢ L"d4[e`'-zMGĂ;H X2xdR3Ŏ[n[{C@‡Dhز'bGZMj>WM!%^DD!E2֑ EIrc.{͹v]Co=Baޣb*fQ2MH4)8K87aéDadtmb ]=Dj(7VsuSAwk,[/Ϻlq/6&E |F`P";9ߡHp^G,ZBpDȲ<Z`x`Кs18pjI'*`G&N+N u9]FӍ4Z،--?2y_Us"nh5v52#`fq*oH3)ZZƴ(QʺJ-mG!})jڲ$KMxVIj$v>P3.ƻ\@r7 Ibwwq[P-9Wፔ~!#= |}1sֿ+)ჩٕ2CYk']R~1y-/ X. Xv5\~a6YT]#n^SSNqu9Ā]1I[FOW{\=pt.;+ No4[`9U|,9%Z$YJG ^|V^Q)=͌5e7ψ8p%X6iU޾邅gz9c7P2 sSbP4" 1TvIS>[m鎊tp+ &>iE %MFdZ@r8D B8K9܊ q+.o,f:_N@~|VRe/W_e|&7^O[·|ip%ty'hp9 @ 2Y8%x ,55oƻDi-C\ZFGoҘL3dy㬉|{8>cӵ5jK@ C4O$u^y(-ߏ>/HCD#9(0[JJWшs p$'RZ{g6FS$!JuIeHz.bhp\n$`6jKF)W~MWXJ~~7Odp4;TWiÔX^]mO^)TSRdKJ#-A@%o 9چ?BhlPEu9 BTǨ%@ (mT1HL451)LCynJ.(@mrր5Il666[na >]? F_G4-h#]EѪT,dQ2RZ9Jf3puB%j@2> dZ*D<~$AFT:RDWxN8(cjLN{xt]l l^~|kLdYC-]Tx梽hs)(*|9NP/I^_@ÃXuRD-x~ZSNuZp+ I(2JR9E@!DR9l :%nU J&GJcXؘd eX([, Glс"B Bd@J m|y)@|P#BѠcr6Wې=eN *mGI2Wanl: a<nL:ڪaV-j v#AŹVPGj/D KE$&b$Hh!#3<'AT5!/Yd8$1'CBx4ZacُQ?߈x4ؔ|uÈ[DlqK&M9HL$Y-U"B'Nq?Tm$+ֻE%8Kg\p^s\+jђf\U{|M,Muռ1:ζ7Ie௛Y,H 0v]\6_.~ut%$;_UHy8/jR ifpt^ ŋŋͪo\A@׵moV9[Z" F[x /Ã];Н*uH~^\[%J~<=٨|vo7 '),ӼYQkUmB ,Z6yZa B" ,ǸYx*pK n+ݨZg;@:vdu[XSCROPFy%;H'%*WrvyLC] uz"ɖ cgDsqK|^]o.6ʏV󲕷шzٽ8Lr\d 瀦jd0vFRNC5|OX3`MP@CB{QRp;i~qW%x- h_'? 'wK&rqӇ_/8s7 )V T{7Eo>,{"GK6`{?̚mx{IŊK+gV_zʙ-%ԛ( Z{:r?}1\mw* š[ꪡ ɸpL5cVΣ&Lg5y'ç׫kǽ8rPw!tᨭ;pT| n肆&9JT3;ۮd-ZK&hcNqO uZ7#`~_Bwy3IS,_?7 |?tPX%[._*\Nm |˺RU:me|S7=h. =ν(܎ lp0\\.V_KWo.wW(kꊃNbƺRc9/}ً_ƐWR|z~ʻ~n9Dw&!r0$7g7#_V_l\E(wAUϫ^,zYoݙ=q(ޟ^ז޹eSd d?~𸛁7^_;emч} ߺba-v~ؾJWWWW?u_}Co_gʊ_ysbj}^KW>|=\^\+|}Ax)7JkmoHG%vb9U<*вS+ 7lk_aoCC^־lZ}Cәخi)Xp8qC|})1'm_k'U~pXq 1HTz%L=DJzSթU-ƫYm _(9]{ҏw#Sۏv3|_d,~0$sϿ12{v} ~9<޳%ȿzqkKX~ɷ_77kamap.rۮ~1nZ,?t87ণn~X\$t[tlu,z>+gB8IC%:2:u^:MweNA%^UɧeoH[]vb9B%Nu}#Zo:@ekB3M p,;ɷ[0” h4wu ֎d\|Gu=pqptA\-!e_^ؐqaP"6UJ&k1n!/U@k1xYZ& ,<eכ6\WqOZ{O8ZmU%ڴte"tevz\HWuŸ`cu] }|t5t(]iP.Z$ EW3 3"pg+5(u] Z]t5]4ިt%!]1nPDWBוPb+$p? f4\fJhDW e+6] nijr`FG :FܐtJEW3ԕ7662pF!EJhHo] ejDGWTP7qQh1g7@ݾ8rZe&MNL /1+K OGvˮ'>~* ԃ(\fZ]vGQ:L,EWv=XMHWl6b\QT (S[b(z]iRlg+ƥGMʷv=7NJf+$3G+Ƶr(bFj2DVtD!]1zE4=RSjmf+$mFb`|+uL uŔ+rcN9Re+MtŴוPUt5C]9i#?YCyrұw+A28G]yB!Z-K,'jܸ=6 cPf\o}.Z(S(Q,5֔͞M=#.q !IqF - kKJʖE2R+_Q*ݙaաNv)},Jk iˮ4sF58ܩ+D(mbi(BWTtk׃$ $FW;uN8ZSוP:,t[ וך\t%Sj2 +vhѕl+u!u] epEW3ԕUV0JpEWB;U(J`p"@یt%.vV+L u@;|d3ZJ~0(>]PWM GpltŸ Ti( ]QWdth̚t0Nan:}e-i6Z6=ϋ~ݳrg۳qՋ_w]-_p/.墨l.j/.*_~݃[z, {y*@.:\D[FmDiT~{0p>G&UeH,Е+ڵAw:#]vz* נEWLkM]W2=jǩIWf+ 0J( ]PW&hHWt6ܩ7Z u] ejOEW9EW9nc *5*,,kJ/'^"gB ktХ }C@>g$ ZJPL--ޟ(a&UtDh&42$VXG]@F{=d+5*+5J(f+M.Q^lt%+*9gڲG|tNN_W<.(\TviqDWsԕq$PX& ] .f3کHQPt5C]YQk+&g3δ0u] %b uEi0L .f3ZkSוP9B:8`FTbѕPZSt5C]y>zԩuOx\%=l<`FgNp i= ʲ":KM@CFb`0:] &Z|T)e *lD43}ָ(`k=.H;hTc(׎H%Ptkkelt+i=u%jY銁lt%Z+>u] h)z]Д(`C6\sѕ+f+ 3kѕGvDj>" +g+:qNK^WBZ✢G_zt(lt%#hC u*'] GWGʠP2  SS'k-q7D= hj8t赇i2LZ>3av,3*˴Yz9 QTUʦNL:v(l>v%TFS Q9fjqAa.S Bk 2dڮ:Q茝jי#quS纊ӤtḯZ-ڥaje+SՐ{-]PW85}8`Q +DVjBUFb` ] VJhM]WBIXt5C]f#]q ׹i tJ,ue-8v6M .b.ZKJ(*]GWDHW vtDqejrrZd`d+u҃-<[[LHTZ"wb*ͦ#]Qv&4ѴZֻ5͔!E#.T6co&V]=]]'Y2ӯF;A$XW1Zb\$+]6HW< ^OwM.Zu%^]AW ѕٻ޶W#o񵻺 dAK?mahQ)8=MRfky$nV yI:u>evD8f;rLsmo"s6}Pζbߖ?0ׇۣFdmxP#3w 2C }㿴\Jڤl(ׇԇvcJiY]!Vns&:G؎pf⠜GT7>8Ma_\B~iY>K۷B>p>f'm/vYěuھ~ˇ 5;:o '4G ܢv?`ZfiaQ>`n>c{z#/:}Wnosiݦ>^#R`y>=˱%͡6F.,UΨ(GB"W(S!;|mićGo^wkWWO| B~鞨(=pW9&kt8.[JFS1X,׃>?)3fTR1FSjSUjt78wK塟07dI˓ ݸR+UE@5בblҩL>V Ɯgr t`wL"TrZ*\ #s#a;^)Ifvh1hrY}p}E5K[[jn\u !qm#NM#reo TgXƮ#D3$fE̍E^3bZ䔔ջ# ф`:3C&fJcw ۽Vʰ6&]J'M9w,F!e@.\YeMK>wEa AQ! AuewS$Ƕ\~=d|Ht#őYnj(`ɘu!g5} ԜΛhdU!ZtM%%j(%%QjnbP9\VTp߬#ɇ%]DDNE[RHqu~0Z#M‚ڈj1*0QYDU5` g\mFO!jTg"ȑ9j],XU!kՕkʧMpS7B cfC),X 8̒QB4BFE׆R3uiLZ%z,/V́ ]d01Ϩ- OYbbm>)h-i0mLWjAĹU X.:@2yĽ͆JgǚȭLt%ρE"n%Q`(.L-#eC;u{|KxٴP]7.޻`(JǍYjXkY#*T.JvTZ[ŨLN=R}A1cF|*VLKD*1+5Kl`W&4:΍tL`AVPꛢDd(MJt*TS!z@eΦ,X22*zbrze ++eq2F[@h ` ePDdBEA31ۢU:B:oY{$,X,x50Mb10pc_Jr|BZ&%@uPywDP-:J`"gIW=$B+ut2: Ρ&y֣⣂E 1ڃVsH&52*.2pȚ[CJ%$0AcN1֔1(28hGGUS U_"bygUcۆnZaBTD:AZx _3X-D{Th Ŋ޲3AKgy@X΀`8)YCqy6sTI5NfWbd@BUq@ơ#bguEתEQeXJ cF8P#YfJm;Qb5\,ԚϽ?h},3GYLg&)&P:nڀJVdǥx/GtYQ6HªL$` f}AO3Ш\L CjGih\ƹo>|ͯW]S,|2|uFgj`<6)c;Vlfѓ2Euh0Nӿ ;-,V ӚZS{Ihd057#{(g˄̯o:̈=X)s8ݐh!/Q[tؚb tyBŹEQUfRA4K̨'ƝK;H]qh`cW,XB|ݝQDqmb}rӐp6r%'y/eU JKO޶8PZ+>BKr$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@pB}@(* H|@$> H|@$> H|@$> H|@$> H|@L}@T|B> MHh5'hz>wXQ$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@'.h\E\OOwJzF> $> H|@$> H|@$> H|@$> H|@$> H|@$> H|@z.>7m5/\cvϟKkuu(#, nNɶTtlKؖ%!P~kǞ,zy~"-d&UO`C3%vQY,vTr"7=QuTBɶ8nEIJXpamoqQq6G,T: YI*JVvXZ|j2StSLOvmעcDp]Ys#9r+ y d؎ ;b;-P$D*yH[ Uė-YB p;a1%ߌiPZo -TM \]ԱVquB0" a1JJv? WG~9sHoWQ -Wq;qu/ߌ!q ioF\rqԞ=_P):qU)qUVUxŕ(=DWP\a !qs!oF\rqU DՉw( B0WH.VU (oWJ)Q\i5%S{!͈+$P{+P+Zoj/T*ۉw( RWHb fLs#]\*ۖՉ+ˋ,qUȽ&qUۮJT'ޣFfKw}T,Q #mM=vTeh0S9},Ye^<Z[zצּhdR{|-6,@-)iKtFFQ]VEYZVz6W5':~pQki|I%=r/t?U^\ܖO5X R((&9f^ʹ4"ՇEN|?TDDҺYJ?[~8~Z/nya ;.7d?Sl"(WKhYPUzQ~s/D5 &{d^dv,e,Y b˨\bJA*, r\{DAz-WV{erʦ(e|1^’uiZk^4x!MfAJu.O9NV4y`أYja1ww}=wӇƧ|pKNǛ'ZjɮtȜVt9sop4xt>ucG.\0ᾉSutٴ;P]( ;t-/壔X y:gdJ%V )Qjd45 ImY~^R,_>cut>?lzLFn+ه}l{wσ-*Rѭ;nZR5Kziр,Gۖ^X1=_.>qizP`2uvM$-lT+ewkllQE_.z iWLnMAy?s pƄL8N#ɞa#0ܙ_nwj] ~9.K23dsM'-xZh<6IH32kQWIɥUZodPLx1 ic.{x޴o1rĵ;#\Iq LR bw Tv)r}4ZRO6[فܮ|A2+)9 z7ї_~x [xzxʤ+!?[QCvFw/aoEAFp*aeBUU ' yL7|Lc+# nSDցSט#_;ȝMȺU잺Ev'䢱mrhd6z8=.l=iVH#|>̫'{6r~;[\mX=X}u>37KFsV86Dݱ!sPpu;LgijzJD"}n#Пـ-*$,TNd`Lng2ٖtLn'>3evDK&7)*5V*Ax D׊Iʐl](7q`%#S tf (TZΜiZMn?w.[si q~k˻j,!viGv0 wu3sALWī@*>P<jMV«6{oz$2A"FB4ZD?6d;4eDG-ADkeddGF dGj;>鉝w|-YiL) Bu6LhQ99MJ ަ$bҗ;-#fpɽ( (=2eG 9h;<9+je 90t] Kk-r1D:wr~Vsq8>?l TDˠE _epF˘ 2`&gٗm`2132 PNidtCaWHkL a&nLD] "G2Gu3atk(-3)ic.&7ͺXE]⿭*2˲=Hot]8-ׯa »m#_qԪ+հ?JӲf|[۞m]x<8/K(+ 86TS `ltD:GaMIf$x❷PsYԓhy;QUDݐe6xsΆ"PLв83>jI!%F;mg;Y\ ġn Am{402sLpJ؄K*GyN:夌4g1`# xiԖi퉶UL;>xFrl[we1چ;ѕlDϙ%.L[>oJr9dM)8)ew 3scwzWcAEAӻ6xUdA]dQ`2D^vyv)d"4 d42hGE_X昒qsZhUf þѰ mn-yI $E,5 +De0^%1+D6s]| A:Nipn@6H9b:#[,DO9(JoXVA84ȥ=`o^ۡ2dB]h1k*:g^`$ϽEAZ#S[Qb)Ό2P<ڠ"P2%%2}W㻵yZWܲǷ[vŋO:<M@2=.u#epK g>? ޻)[}7ݣѸon;suY aqqx\>aeLd"Ge{=Y Sԏ&ǰDGGCLէd6.=0ߥi4 + IǹZcB)9`TAzBџG?G_*8x+~$y?j9!< Tz& )B=Dq|~ޏ K'*Yk"($|N^d͕ތ0gѷk}{Ǚ~.G71N+`(u18$>Y9 *_ek"Tx7p5~EL,0VԸl3o;·'ayNҔB]:I" c*3*59r}(0R(qUQsb|pC|u%fPu:?a皜WU[ [ǯ.fޅO(gh03eѯ]R^<^]6߯h4 PK;bۡYtK3gC͢p,6o&9<#Ӵy{ { EI[F>:66zn-h5'>2I8?\󀚳6ӆ6#PTh- Cܙwu,9 6<$WNo_x)ma kUN)ڄrY-"p1z,(R ׎Cڱ#EB ǯo0K`$jV3 5JSfu/z /7!sV@90䓲Jo~Wyk2'[cns:Uw( eɳ.LC9Y`Kk%$^m D05u)QHBÃwϹL*&&̓VOs$[#gKa; e>DOOºz8Mү_bYmzfZVFt,*{LirNPQ62ꠓU)E WΰJˆŗx>(4+"eb6s*1h,]lLV`TDlx(Z͛;A]f(`ܓ4L3Gnf?nrG :zl䌕>zÅ&1P]"ƘLt@3Yؿ۟fiڪLuQ͹V47ͼV Xhخ@ʄBHb Qn%41$d3V"23 JI@6F *IQpIQJpobRnp%݆_7u,g]?g *tzޘ˪2D'ϥۤ:Va'hNE/ś=^iǣ>HRSUU(0E<=(9 D8{/5n5)%7fqk#?c{y s>?׫#6kmY8s0͎OҴsoIiro|=< ~3;@GcO A?ŷõ!'"ݠqףG=b.إkap"2 z*QP`oh0|E] Ao8#BJ)])nPyVeڂa#2"^2'r)`CuJtKl{vO0/zQ9GDcW 4Sɲ, S(3(ުڨSr))!jS^P# lJ \:S#u;˝^iTjckls+dFȁ e2jḱǔ| V < oV$D1ګ$њI0*lCkIUL~ $q FTyĄ5r6 j:94nbWZ}6N.5w{ pi])|K@#5ު9k1ZߗVQy  $  0Y;%2NiVUdTFHSГ FJYt))*.HUZ8XZBazXT韞f9;X7GUUvBїA~-5!yZ82ؐCm^ʤ '!j-h4)޵W {.2 a+JDll;blf ԪR&A5r6# 㒉+wu˨;4[9PPI IvŤϢ Jf9Mցj;A0haJ;P$`!fH&kb"AĀP&Cc0 ٌQ?߈D`Dl?eD4"vx+&h!RYz9`\mZ+v%WZ a,"GrhWHTWHzD2R n`) *v^&\)KtTGjEe~WY9wgwد;LDE{Ed4i`rȒ$=$RkCFqr{>f(jv/*7~"D"!=`YAaZ[-5pTMZJUN!- t3<4nk;≦D VTDn)B -O'uOnL(:k6J96˙5(W?CQ})35&o(F+6sK*.Yeu7m0e1V)C efEVZ%y鸵:sBMZ~'` %Nƃ 65tlcXpI`2Q,-Yb461eu(ػFn$W>6,vb"g#ˎ%dzdIӲۙ2Tuwç8 #,TuAg/1![:gE97Yս?RB Lc# ̇k_0i_ )]Ι}`AMAo( \Ui;\U)vp3ՋvUW[*+JYm{CpU~`[WpUKd*%p-Z~Cp#7WU\ \Ui_b0)J_y2GEuwOr:n1(=^peZDWcϡ?r,}cEDlƑ pdۧr>϶;Raݡ}g82)^Ay|,P@#׵*HJ$8kj-V1)/;( ^B |\pQ) ei"Y7v&ΞVhڣaJ?ym\~-G?߶pWԌmMr}tD GG+{tVK3 %`D2dY$hɶ \T0BEV_&u@ܓm/busjYZ$ZFXZLX8:yg,[Ftٰ[`_,=;6ݚun\%i8_@6쵼]OWtr~(ծ4H+/G:F_S%eFѿE҇zQmHgZuߤ0=/ :g:Z$>2H]hnG?]\e~[1iC0˧JR[TOxݽZ PB݊w(*4_錷<9 cGI>BHBrme6k'ög~W*\q"O#\b3@U=<}ڭ}?,o]e/.vjljD n ,*iS,ʄѹ$Ul$<ͳ۾L'~p4ml_rI 9OnN: qv O\K3*`K9o. F!6N &*ZSmOkڞ_o 9,ޱq1 =wyh) } q}UvW٦'ug1Kx1ZgHR wSF  1YvJ"3D6A@:eD[d|0""uN{d43qIw7:y73 Y7%pyd /.~YS7rQl%Z=OJXX`!P&vBDh͇x;1{=,/ @oVKo ``ҭFۺ5;߼c}͢_;plt2S x]aoN495^R_ 5|>l>DQF] T*&&{Y0 g2Shs9].:9SDP1dGP  ^Gc4g'dQ Ps a``"3$mjt2ȞR{k [* 9MFſtn3kX:P|ʚYYC t?6+UMFB0QSVԱv&ziik$Us_ε:ӓ#ϵ\Km;}nO$N0>y}j k뒋ϞY7ӥjoqZ7G" o[doŎ]Ad$~U':@iXb uBTbAqГKNWJ%L 6' 52v&W ;ӌctf“bJzoʌ=a3i6n 0|_MgsV蔣 FuV " _ֽ8%YP@.*1CgʩH Ş6`4HVz};|,rɺ3qGx35/ݙvڶcԶj v3aht`Kl⟑KFkdsR)ml`ZCffXT5E22Kc"D*Db:ۧxؙ8a/"+0ҏcun@wi 9c*"giLȬHk͞DKD9_ꣅN;wc{ ;,Rq&RL X]JIK*XTAuG6̯:2.Nq)ٙc\.mV.Zb@N: Oh%PQ5X'D=)p/xؙvB.pxx[ּܼ#jcui]5OhH_)ھr,0 U\JW*}]0tr;kEͭ^ty*Lrcѓ515; 4 6hjϡ% 8m"o^y޶[QODyGrZ]i|6 ?u Deu 2G""AmғD1B)=PE,ba+a ."|ǁ3o.f>%in?\k缇]骶ociݓj5hyZN8.43fU1zM6d yjv;]E'zP.}(}:ޞKAz_[>KgoBmj38`y;AG#0_$5e5vƘ #ƈ&i%71 KXB(S>IO:8d$.t@K.7>iRAg< AIB LQAPD-3:% DRsd.#dJAT:rB](MLr.:k;PX,l!IFJ>eTj?$E9(QdjD XXNa,W0W '[&'<yN:BdL` 0lլU_'S57UXD_1۰3[I_  W ́w֟9fG썠;`ud& #`"uMSfܐY5ηk3ǎ7>·~8 uU.reV :/{KD`V!ڻ̄&Q֘ZokwUClӥƢȡ`]io&: e7U,{Q_4>OF3nxOoסwn^}arisT7N۪qxw^\m#nGI/8R ׯS-bϛGwS^b+^sUyⱭwwc+Q=P7{Tuk.[J| jU 5%JTvpdW#4UN`YdJmYfn&%Y[d3D.-%Do:)!Qy6b PjHAo>>|[d7LP7y+PNMLrmT;yy RK&Js-__y0WOQ.ZUQ1!j?|Wͽ-ŸZJ  PIaKFhLW,*Oc*w?{6߃qh2 9oUK?@֞KwpAȝMot6TfumrR=˷'M|uW_nW^d̢kvg.W6|9z_p/JXypwsc姻y.fFtXh! 6])fSщQIxlcUI*a+g* TVh]ENӚ[~6~ZCH[ BI dt`-u鋦.|w3cOuݼlg6X뫋5[(֏n/~_q=I[14EZeŧHi (g[{@T5!C `sV芥$ ynOkD;*SfZ&xsΚ)!W*Y4QIfc@Q;vjѧ8%5"-Aj:!Ֆ!dd~eG U|0Kcҡ2S)g`-HxztiO'[z0eU]#:fFpq3w]V+r9dXe9B%mW\Kڸ6 ؽy٠y]˧x y #dD09i2DpNz OBȢn6l 5x\T^Z"g9d +<><3R!D׽X8a߼9pp M _y&g :/ 'p7)eژHdUN`q_c˦X ڤc)&2ΐM: mVz3вD.,&}ҁ;9OV?˥dւ &`r͜$~N"JT^ KeTK^ C/h.W>Ƥ-98Q=ڀFty$B#Az(N٩i-6\luN L>R{xBCP!_Ց"-$e3aqW6h'hm)ft:G#uCYF0QV/$AGp [/,PAA8$Z$y{0kX' 1Sf6N8@?G6Nm0w`pakGG~׉6NS`H#JVrO`H,RUAvpEmJ5&ɹs=q1G4ƒ;.{~9_0˜_ra7i}饧hk"TJG3: Kx&YDa$Al=AnuJoƄv6tk_-f_7+S֮Q/ wXa:CCz(;λkm/dMx'^ xv[Љj<Kw㞢A(!$``V  twrLd$rK#&ܺ]:Q"aH$BRyl%$4 /׌]NHs&]_bD&eUi[eʇUںq?zףKV|Ӈw;W駿ݝ,^Lrhݹ*y .\& ep%g 'R7poҕeTev֋Ģ~\!|;<nEE"tFSItTd&qA#?TyRܪ BHxc~Vbib\~u$߲>U:ߔT{#K^ 2O/;<fbSkI+Y[Bo>E]S]{gl@ ,<\L]#,\gǯ;]oksz7QL<"|?obfWmVb 7kv;c㩚%ty=-;.,E:VՕ]WWf#7J ![?L/j6XrJ+V Qh !&cאY1ZFyUbaY=eԴU+pQزH\T&"hd^D$w[Ʀ&8(okw\}]ӏ d,S|#|Y)ÐNόTѠKxͰ6q6þIf5\oͤF==*= "zVEdsVL2˿L+n_NzmhHV*OnR*x!ʠ=&1#::,bQ0@5w8(&76H3dјT,jBbeQl+cZJs,@[Ij`mRܱŜEg&6k +phY"lg'I_tMs5HzbIi Jc2b>Efٻ߶$}=,~? pwٹ;M\0ik,SQ jR$Edd;bYlVUJF05 D)oӢro#On(b5W {ˆ{(^L03RxiLa2P؝0])"ddl\ !c/؄Πx$wR݁檀%rV:wr.`9OƣAl)<@rFϵ,w𨆃i ov G&B؇Z/v;6:v4Y;PUlՖ;srUn6[cR;ĬW逰 ߜ <XecP-rI6]`7Jې I:ƥԿY-s}O> 癉u=HV HLMjR砵b%ڜE55hvJ;HuH%D|*d `;k sRnZI9#44 C!DsN\=sb~r{ءΉEKO;#2J]PJixk08u%CJ WxYhr 8M(rn>xzWEqGkZ  hl}klwցΗ4)5xB$lamU){﹫{tx偭kA#ALBϖ97vX7Jd&΍&$gX\S)sa[p?;ZcK.FqyP%鬙#GJLj&H:ԥ"땀Oع77]~׭|nr=Wءyd3XndžX R[qE}j%THE Z(,Jm9ịUh4N13*h2:ڀmTy0XJBxJz"gY${uIl ݘ_{Y{m+:H%%E5 z9$DUOmj)b^-QIrb68k@U&$ZqD9 fykw1^GZF冰\]D[Ѷ5m3un+m0Ͷ3 d J8Қ5Gaen.aAۈr0(Eb![cI"kCb=b-Fu^GpO~"JI2@X"C#9Gu2B5bمr=9kU ôGy 9l'vs|X^z_9`>GPڊ7 g+IùIɑT6J!۠LnLu()(= bύس#7tؙ<F#RD `$HОHob&ۭ2v0ϑَڛNUv@byg-R)j Ia2r{ GD[7IBZGGtz*ڥIsqr}8u=Caޓ_q1\ B(r,WiA& ͰHQMQ XIBi'E}z*Xg1xp6`s?W13 lӎ_mkaqQbbHuSν6`cFņi,{ sTTN#D#$^ٸS b8bP1iֈ2ʰQ:p[%c3L+&ݤ<,Yydc"k/AoqNJ7.Ŝ(qm$F.4mkysoJK@zkë4*#6O"7P}y;PrClĚ2r xʐ "meҘtzJ)}xZR=QIb h&vTG0! J A󝥌==el8nTyy6d>3u-3 >?$ !;}ݛ.QpTEJ~fԻ5Ui avmQ,a/#ؙxнϊп%labX_Tb7Q5t^{4Nn8Hw~&7K`,H02\`y1F{\d>08{v̢w3}HϻiTSZ_f0G)˻Ͼ.Yhn9H6˭'X? :QȐKU$ħaDGXVYiKk8Syxcv(UQ1>Ds7X?H+P[ } l_, vﶆw7E^-3DU9!QI,U,("{;FIc/euK#?Q,KmQ~5<q8 m̫[߰ѼӮ}盈h9u+7b|ƳK-."UaaH6S%T  .:r[I"5&t휩'<__hK-D}Ef:eR t`h5-ÌR!1{4ڇH9UHbdSz(ā.^e܍ZKWu^k}*ĺ<-“U/Ɗjh|]D+qT`E_P 9Vf#KKil$*l|m6$#J.{>P>*?ZV3IZHU9DCPm7OQ,SweQV?`]ף`u9ZN! Z_X[1#>lߒ|ur6ke-e` 0GM,+KgsFV;||^pv>q:xe47:Ru""ufIF8Xo>5Ÿjds_$gT\yisFQDfE<);"{wXtOZ~,ZmͶُL۽eNi hg;hȽ:z'%qjuTSRb&y$,ViǬQFYJy>%loP=^#Z> lm6=Hۨ>(A4nJ.6FŲ)4}}YCkY3r#$$3Q#v$80ɠE=J Cm4om [1|ჼd[Yo3ב H 1Nffi%1 6 $'Acb(`ܫP4 @_Xw 8mf[E*"1CzTaK S˫ZU&I2ΰs/`>)"g#tv%3S>n@O%P0 tGtb S ˬ%W%Ĝ\ .kVLz"U e Fć牣G*{nE6rdP&^aTb)BΔUՀ۟y2m&*\X a]ܚ,00ow3k9IpiD~v>_y _8x/ y_/{*՞jZJ[MXaUQ 17@I&E:X%+v5D]`VAh|ޞ ϧh x 3=,' $R+ܖ)IU*JJYln|R&Ğ&gmP{&r> 1`Sdӡ'8N\sf,Z~!EJ|P"Iր`iѝuiThBzmgo^E6ԒM.ƫwK,v.`zo4ǚY&JV_)-86K IYH%HJ#-A@%taBÅPBv2 \W \0*QYíN*d˴5q6 dgABFH@)EoP[g*x9DZ9f뱜k'G?~ON goupqI˵)oj/qZ??/xA'eIiG繦@uZdhV)"C!DR9l :%*I%h`Tؚ8둱=bl e0?xɶgv\7:j7pR\8HA WiOr/ED(t &n[g=bA?̸ǂڭiǮZFmաv`Wj<$[RXA!hVeI)N蹚u4 \xTi22C 2 &DsyDc̗!u!@<:Q=5qaO"30 ""nu"xD$7A|"12)p2:dEBT=q g)*"MIΝXj*Ô,eqy h*I3P* 9v~kl͒+-[8T8;ִdW\4-pŕiN$!$ a[Ų-~p4%8J'/ "z.ifxbwU$sfJ+AZZ 5^)pBݱlI<M\ǓZ!4TJTxjY꒹L\C@ M2*{}"eKǽ%Ɛ Dih!蠵9pm?hMm K{ƢDeJL`.x9/pjH4Hc)ATp p!e]VOm4}iǢ52*ACi0Z&`&nQ2JIxyGZ[.x5l8>زнp>C>NX=ӝ#<9%w4pMQsTj>mdUH}2ҜUS:{eb"TL˪O8zٻM~XJn/ݿt;t ՟m V() gȋ]5m:%* *7F+j̱"0K}8|^ ],r*{/./qM yQ:N؇}/zn~>p, ~qU X5xXk.>sP=uK˹턨LTgb*kd56!kzDJMތг| j ʦeVZ2ܓaPq)%şSw.q43d 2LJO?{k}pwUB72Utow,󏱞 ՚Z.~\JxC~裢h.l\R.R Sָ ؅Hў~7g&(As+\ɛ(ךY +XO;mMf<.׃~?hg9wDJ!W#KcC*s]gjt3nbz+\!Z^ 4謖UнF3~ധ_Z؉r"x*Li2Nt;q2鎥 xbwTE.2Qr4(Sp1T7Spq@ `7ݩ! SC"Oo903l 2p1tkox/@&t/ѫC/վ+~nc%pZoGËO?+]ѭۛC{(#$ȒzJl,9ΠHUjwzp7'^R,~%)~C7wE3,5!)`~0hTK%kUE=T]VwG{3gюkzb(V1+B(_/o{%NMf8}39J]٫\𼨊4__Gyfbp;q^V{O,[єvܞ@] n+E t]^?i4~v_ a[W[upGk:Y}WVf!,$MݮG=7^yA3-Wp0XCwD78gMk:n<ekwT7++9yIM8rSގ/$Eݼa':/}qi[NS7 }?p3b@ ׋-HA tu*LaeiL`:XHh꺍m#XǵZNA[%Js)jӔyBYS7\. G(pY7MsM*TG#u` `AItm yj;Y[⬿11fBgg~cj$:xr+S;#iySxΒ+y(Nø2O 7wr~,5L"kXNMh)5і,EL#!꒒#-B`;PNOd] !Z%嬱:oЖpi[ vS0`D(=1qx@Hu )lklTa/o/_Т3(|]}1 [~'_ϜJ^l}'̋-D@KEATBM̜3SQr\M:b1U*&hq>g 3), @U sJXZ6YoSuu;QV9[ LeqTwKwMSEY-L+<'朝LjW>47bput|N% 5ԕ9a%GUڛ3|JWQ#)f>p’6̠&/ Hz N$.Iu 18{??mtuS 9GeTWJLwfl;u*3m$;N3 tG );`&tgkv %U4irkxFdajk.5>b@B Qn)9s,.?(,:zʍRRڝsJmBXUtL%%U|-6]%Sǽ)*rݻW 30>R 9q&)Ym,VaN|2,jRo\(EEVLЁ{̤尃:5ø@PNCB~wxrbϗkRsc|sE_coOhN7Ň=Bb{蓤CS七;ԘQ{Y"6M^*]45ʶLQLqtGmp'eO(>kT92e)t+t甒s?o2f$&+XS&ISXX)b[Q|#0Q?`M+6s;DL{T{? HԢZpr~5~㿫!N$s,]b RDKpV2(cDM_r|?Uv 97F*+TqٷM HV4`Hi+/-wh?|qo݃rp:/}423 9wOU~WCx˃W}&Q:U51<ŏ^728+N<* >WhV7K-5=/֐>sY4=6JNTv#+dsnC3z`T i,#΋rsLL LC6JAx`OqtyI>nZ۶C5r* WM$bz,FKì48H$bɮn-n~3?`sG1: +D6 C9Gmjݎ#fܘ (,zW֋WK:*]Ss>DnUfܭy]L=T\V|?3 qnRo]^h(ϽxPt[˸8A@[y`z6g/8ȇ6o.+jkmWyrQa v<-Cyu |Hl?۷ uy6|n0D\`n+wȏ1K"}MNNaD\8U;Yzw]~֏n.ÅRRݻpŘ;[~.?ϝ`oty3Zs#Wvf`1ggơgPRnz>;3ho 5XU\M~* $Yt'+>O826{4β|"z)VUt:nxo}@*? ?| )'Te7%+>]SΫ_DDM"8IR;fR_wK ~o~|9\cq?􇝂o9 z׎>|rrw<# #eUNkm,'#fkY( U BmYnʱ밡] V5Σ8 U*K ɭpJW(U5)YB*^+P.WwZm 1k0K$.Ư p4Y q=o@i=|=ҬHK(Y3ݸ urԉ2G,;:13Q';Zo#1)Gq 2^"f<9FqA),NJ 7Ř 2Ye&;nD"^˨'Z:BeX"} uZr=Є9bz{pd$1e PV#G8#0U$ g;?ۦhQyafj>f& URq_W=8<Wf%KV}Vڱ&%eroKusisus7y$ (F@rldHQѢ%yAQˆZu7K-bxA 7o$}o}).nn>^_OϷu GI]5v^>>1şWxkLavk/`La ^U1qgK%Wޚ^W,+c a 6^X0 AЎD`ʔGkYf4ꍷ0$9sNΓ:Q:!Z6$* iǺ,Rh|i-۾URwkTW{k>j҅VZMp4˸:Y,dYoky[ފ '+MD @]7P ş+Qc-[2y]~4j<wj8ݙP]͜;nm5#`T0pn/Ur (70( ~s1}9ijyͷmimkZqߣi\%k kE|mBR4mdZ6;)aP:!:O&RE 1YMD0& H:&dXDes9F{|h[BL&h) 1' hc>zN֋qU 2^#E~`0p_#d΁)~NΪfKkYncbY< Ra6,1юvJsd#R*횷NMH 8aMp&:Ѧčqg GX #xmcdDX(k\D !ceK#64qaL!Jo&D>ofEtʐStt&#i템U1oQӭ[UuRbM#`cakS{Ƚ N=QFa\dSMTBKh<ɫTeԱ5\ZGݓ:Pk4nt,Dj%`:N,A^"a+S٭ZVv $A:MN+`t\q&̌'n ]]WS]'>>WG#whن?ɾ:1g|θA ,/T#fP#Υi}1\z g⑋l*)tV4Nwe?7?Y=,gX]/mP],DiȈѺR#^ԕV%))#+ nJz6{=)cwJ$KP}.f !orWHohg}<e%T0FATGL؊ %u^,gR+AL-GUm71uɩ]ƃʘ}XӶ#ζf˜oٶX:M J3+e{K߬Vy_ZTtE[rQTz ު*|JAV;ֽ1;P9r]'뉹q!D.E8G L[ahꂂ2 R23$)Kz:`o$6Mh -Kn:;@ n$(FsC`Rb9;7riҁ9tNÒſҬ!4ENgzɑ_12; }^f1l;Sd뇙2S"$)g7PT&`v(#2 B!Jr +GʿDA%TRH8KO%NŶFm/^fK2$ Rũl|7Uuv\}:;;Iy{ʴ>׊0N'ŎѥPFsST&r3HIƐ+d1"XI$AZ! ȱV "Q3<鬆ydFW uV+X?皍L{fMd:<;*s^G}ʢs0s+)xbl53v]/'wMȭն#.ˠV>Qz>:#bhN%13T jQF\fL&4J"br}`H4MɲBY n!xk 5#//,ďfX=$rL/n)$I"h)U$.R`v ѶCv5k})ZA#'CJ[գ埿*ezZ{|hǿ@M)N$Fx,xN2Y %y&C.؈%$̻Z%Oچ5(Zc_-id$S֎0:ڡ$Wdzot_h%cࣉ3DE d)Gc +P{:iCv(Bu[ơ"Qa.4W$lt QkC6nwH!:e(&5LA䪨J6ݾWfa(PPE巣^_,Pc ^+)sP>בŲ!#c~{Mxs454JyH3;$Tܑ"@ŘLD *,TLcp^#PbE2\cgZ!xgRg5̎ebPtV+_[ z]Պ` γ1KӲ,ϋ" KQ+-iWt[).b\^_~\Whڶk50"鴨[R,W_{SNU{=+>|_T$6# mYIF@|IOgl _iqUJ`Z$yZ(enTSe/zCVID(GuJE/[Ѯkd SRVZ5ȷdI`^>د@Rb=i1X_,s;k7vzp'/exz)HgVJ;͉I#ogɅ0$D/ӑg0K&իJFRh5FԼCST(94̧/KME‰Nxiѿl ŸϷI@ }>ij3X<W4o'  (?&ȁXW@O˂?s+ z? 3%A6e\wS-^tGZ,){[,ԧ)-.?\xfwi=\IO496ry:/ o|1yvqY]ˑHMԔyXll-˞Koz jWcv bE/ϷzbC&:n.oj[{w`Z4zr7ݴ`wSf˛Z?"-`}?,m]ƢHeXZx{X4Ntr:jF3ʅ3q^WH-.?~@QdJJ1hrD`XN(%ًȭm9l^pkw9qMbH])q^?07?{ 6=đvvׇY*fs \<9+V^e?/M(?oMnQj}#`|HI '^$Y6w+ge[`S/:u瀫\P3|)2.p#s W } q&9́]n0qa(_O7pitM꩗Ђv}ЀG~iȖIZiwZ(!»R!pRs21+e9-XI2> dvT"UJ3w:uOʅ8NpCeA).$H1iT <$H!Q[FQ"#1P@'C嘺u X W tw0 CƤNfkΖ\(1)a α*43]2-VdBt+z ;E@oЬ dSN8) I2reqLufPg vPoO<&^ga{ǬTܜV{τ0ā YU\穓hdzt`gʁI 5=}b1[wpȸc'T#b۞if/+oqx}P9DH5ɞWe'BTQ -./ұz'JBv Xa+`R@/jW8Q߆*&9wQ?͓t?8[?AXu1TVr[>~@7lFm8NXMtq 9lr҄SHEW{(p!Hti=5Q-)ȎSDЅ[-9wqQHkKÕj,-%ɮ $%ns;DP{5RBz{~ԅ*:w;\RxUgnՖ#ns,a?]QC/r%nRyè] .}5o dJE \<ӄ0<넵o <{һȁ^^ |- ֗\&+&%NK@C]SD!1VduS !/+L>iD]ӋUwD>4ˣeE'Nn]TAʽ[UdRWG[oJ>=[VIN>"] REv F!}솮|q?E3e ߑ Edtl1]QV dՈj!8wqlW0y!آ! E(N1k#J}yc,Q| ?Y7)4D,ĊI-, ͑b9L2<~Eƕ*OcC\d6Vi !K) '#9+iU7u{r[FQ1g<)'qM*=n.4l=FqY|ghy8\={-Y^&PT,hrEr=[wͰ$89*dzwއb}-`ز mr{k8ɰ/son}$"TE{ʛx,bGF-Ewgb-s ܊ 2zy'?ʣ'4L>{X* Dh#j!\-uaSf@%Y6$-:ß((us c s-BLs$R>xY+R\4fmb)VB*QplA&{(C|׃Xi`ઈdԧ׬ e 8Mz d(â@A 3QgX3-?z=fc$faX ^) Qq- ȳwmˍW\x0cv*| l,ɖ_mQ(HSζg`s:3Yt߂,=,{"C2!GeS.S2Mz&l*v㍹sNJ 5$ͅGFﭏ7%vH)3. ͠ẍ́aV)\rZQ8/լUR7ƨ0֩7BZ6VEvƱs/vnbmz^S>K+'Xl^*y}6lm Jr5R`>[dr5n@apRժ1Sa/#P hp͉ZEDћT-~ se"(=?q$2$=9OS Q?7/瑿yOtZ &Udz泐\:g +JM^(DW۫W{/H-i^(Mwi?on5w4_ͦymwڰ`3 6C<}B<5CđƭMl؝k(*k E4bǑ|e'rQ@tME+KC#^y/,t* x֦挰TOI+yp0i"dcR ҽ Ni kbQv0RqS mu-f2.VABHإ=m~s6X f1_bkatgJ\xq`P?4NH[Yq#{)o*S|("`nBADmq[ W5cۓ=98TT#~z)J!jG4t ua:f]h#ƆGC7_,x]~t:N9 m^FU @"U\df8 ysn,6Ċ3-;4PJĪAͦ<΅8(4hkv<$ ^G1- ǂRg~(&{Tͅ*l=\/!l)@"MUt܏klcFٰ&v t֬qcW,4^1\:(hI9a$qi.NS#D]AI ;ybY{/P VHn6!F!㒋vRC'BV{X/v@bCMn7˔6r_7cV['d<䎤+nD<)Mk?' 4nQ4ь6"b!KX;`[q)ohƓNAGg3j!ޟDS3_}DB^D  Æ; \!vIs:ARYQ%ι*H u~ sxAtG^9.Pլ)д(hQ9>%2mi7@RV(|_ӐO:O{kQ,IȒVPg$(cH)=ʣ˒IZ^m6Wµ$ *7X`A4!"$8}JeU5Ŏz]@Z0C  RVd3($eǩ9H] Y(W:1GӔڲ|י }ǚdFe[=F"u *K8g լP6[Odz$["e/DBoCvXU\dؤ 2 ৣt$aiIMoCsHErjWo+} ՗sQXe5-$Bi W&Pbh5[Ƞ5>wI1Wy<ۚʓ##Iw̜q?ϧ2d0/pTFMcլe6N>Fsn ܉ĺScݬڥ2_u0?@ѵ9Âtx 0j-DN5CiҤirkSf~i=[/+L-XXia | (IXu[e0G80x(0%U4-:;n8-:pv.QĢppȹ+n3a)bB=( 2.rT~ҌL$M W6 h. p4kAA]Oˍd`\MI;WAKFD &v{n!\ ra ;㙟Hm)l_jgn/8/Z9Fhp0/Hf9:}kQje%P7 ^pN6phڢ;7WuMIeG*/6wMJx*,2t|mt^5rMҶ@S6ud4Nrd{HTishBE3 VrnU*s0k9`/6?ֽ☘Š7K9]c5k{c!Z<3br Ei Ơ,O,^38ER5v$4i<{;>j8Bpg\7NFyZHjel6Aa%ޫ[6 =Y${AKʹW9;%).sJN<h>g$}x8FIÁ(߼W@Df` 6UiI,Oo z)*+4ENOT0|(`+r@ ~2x܎t KFQ269hs;fO-NY/[/22vy#viI|_g|C nqm#7~>:y m:]ƹS׾q.K+ƻV<:jNZ;?s*yl7l/St+]t-`eX}\^.rT6cB4C˪dfUM[kz8ل5kܒq0kz;NDZf[;Hv9` \ .e]6-l-ŃD`' š5n-NEN֬qښӍ0v'ך5nd^|vK)PV6֟,2Ac>ų[l8t0)w8l삶z>LWN5kBfCoot҂qۮpdr.<{:)a1έ揜g>g'z+@!~79.<+fn_\jsX|7W h_;-0٦Yjg W%)RVou\n_]{o j4joņjt@Ո2Y0 _4(3ϽذaߞIYKx8t#w2+,-c$u׆u ,f33m it1};`06DF\zswe&JOp>V21C+SHM)]|a"qL&25*kI)Re7z "$^aTv!UA_Ȧ{"yX/ r m nLtJa]ÏRj y%w|}7ߎyg}U7%KsI\J{Ty0[fB4EOW#+HBq^c)$ЕyF`DhYxM2%/H+df8 zyXNW|`)bZnG@ D4KnBD!ZdRx -҇AT4಄ ;0%LTl>B;,exWɿD>_*5H(+lB+mE \(H\wz++)=]BQi1빞QJF[7q)BD#na:> Mi;iiy؈5Ӱ#dDsD H CӖ#B{,AU[A>ưVtΞ!2~MR~_"L*eZ8Sp*RjA"(Nt;d fKuR{\MA׎Ǜ 5FWȻk.bڃlJcՏF-h s5l7AZgMDg-V=Mv, (y)˞ݔi)Rca2vWhNԬ-hKZwlloO9cAqB5\{+I;EO9W?ucw3+D{z MyKp ׬qt-IK &>tP_.W;gnAӖC\L^49ŌjJ4³9ٌ}ftSh5'|2_AhxfCBȗh|YW׻e:]-9D8\`єѶ\x:Tlc R#:7\ X\([SI7םx@ hiN{ዩhbgXCOgj5Zoz 0@; v7Q/ ,aW&z+}h>@)c}pp;qiZ(NOa;x")Ab1iH,80}~IV.h16=22#p>*x[.ϗIRHeTpSurޓ6v#W;;kfрNzA|ђWEOғM|RdU:YGE3m}I<p.E/틴ե%ܔZzKwYHk*%j@ǙƓɻz/L="+)G>;c)5Z7]0ۓ8CݕwxFg Lb-d|dzIؗڐ&<;C{!|\Fk*Op1ʩa*(ƀ ӗ8`q`z;bdƣ6&p ܠC+Fke01&r]<͹8.or>9!%eeR*|Ǣzb o?~d05Q8Wc2xG6C4L[IԹ7 xpZh1_n+@@fgK1^:!!PBt ׃_"4?MIrA~r;"}Q_ߠ5?bӤ<G4OwC'=[~\n\n Zb9B\ÀcWF̻/_=7Lu ~Q2U LJS@P))yoZe[7Fh˺ 'UTl [{,-ǩ#yq6PQ6 {w3:r:C LLB=B UѪfT(o确v챬 L}+VF9a5k$I]'pJ!DseQے!kz{t%%:=Z޽~J 0oo EyU0==vF-%J ūq/rw}VBM;fE_hEh"ɪW.i Tog=6S}魂c雛?{^~鸀-mGv*4ഇ /5ԐfãW]7Q@Mo5FUP5>X#t]  hhh܍£g01ȞE2ڍ>tɧ+Ȳa֏ouxs79=S)c8q"\IlYͨb"jήGB\ &7E$u͜MTJ9iz'1ǎcS* ,~4=o36b'!VBBJ:Pf[o^+ӵR~gpwf;~6G!ZmQ6nMۛIIo2Sa z efZI=+_`@gv3#G\@r;˫ls"e]\,He TWㅾ?KW>l5❙ESg;5h2mA7_8AkK1֛IgZN‰FsɉQUw4o@qY˼A0A E.eMDsk>6Ŷ-;2̫+pQ/[v 㗮~=(Dqg\-pwUHJ/<"Xvogl?U*OU1jO}wC-\fI:h[(ktqT Nm;%)­_'ZsЪYmJЊcr6yy65b٩/({S`4!~Ad.%h'5QGe\ozt? QN|aL>;Q^|q= F h W;7($cV0Mtoo1c 2n@ <ĸ=H*?)^߄ 5v|y$)3cEcq8<ϳݢyE;u?.&ax;/ѡC>Q P5e[#\WH$P9s2$˳WrUιL+ZF#l{acgs@SDjCLT2$|r!s@rɞ릈A'+Ǭg2`l<xakH1Er\D Ah,!5II0@Ts|TnErLS_).0`WQު:ZZz16h4PmO˴PpT[Y'))sN57Kέ2!ւHH0nC2# 7j ovZGW)23cI;Α"Do֡]`[DI2m X@kQѦ-f]ށ9Ao+P ge:őLZBD EG]r::KxA(qcrfXoDIJ#TibL )Vqjфsplиi*lB@FVIIM5T9 C $SeN;eחc94%T@ˉoM,`NPܴcvshx2daG0im.59+!ztQ5d]'ԔF_9t+QuR:BV)#6>$h<0#8zBCb ⥓D7 s -mfPKTTjP75k'#,Gs@;:)'JjšYo(yI.Â&EY !5 \m(=6&8i+ >m_+ ׾8A{&)T#cHjtD ֠luF8Pވ`kz4imE]K. z(1I0]ج9\r8Iȶc&,y1yibiCBm8|TBYSV&OpJ-J_郋_-, !@:9cquHDv+I);T JՎ] l}r^`xUlI\5xNi6dƔ.h (.fBiA1f7%["JmD>0~qɍ) AzjcUSM^b/d+˶tyP<}TZQ"(Zwh$d6I@X*Ta yS7&E}`U-\%B͒'SV*ēD'OjI@uzAObl:Mcf#+lɡ+ε!K+Ԋ/8 c|b`ZK J sI4uM @L-*n#If_?դ說f@vqnff)wҘSnAM{#pq;M)˟b7ǜXZ}st%95x's`fKŘ(D7/Je>Ug=֡ Vg8 rW2TEЍTvWMOnV9؈8 xDZd;A3zT׹Qe"zZAPқM+cZP[3UcL[ea*CKnNY7scNn_Ÿ9Ip)toƛ_ՀJ0NVK>O&nU|q!iypYQpN*nZ7-n.9~F07\?r*zc$7SIC{˽xToT5@2)AoyV ,TB(R 4VԖ[JWj*~?Ѕ kFũPAZC:M%ef*.5ZUwn6;2{} H->r<0EXb$\U{ N)[tv|*#kp8U29rVJf)uC HCne7ZU$<܁9RCo=#|RJ"4v<2 *ArHeoCNZ3if+>>@K8[ycQ+(WW0pBR&q#gS1iSkgH7>f{VPWUxeSml{sCYQ2s ^ZKGoݴK7'o#-pּe'5#G̚5!-1e'5$ltM$9"4x+4L#ʃrڜk軳30 9^`m/Whqgc?Yt8(Gvu6Vz r]z7yJB A^1#Pqì 9=z6Wˡ]X.*hhmsA)8rUV+jV dm+/y#VV],Q 2@=?+X\ ӏ=%x}#<)?qSJn{:ߙѧan˳HZ| "T^1{噍R#G!]i@Q﬐ոrr QԓI.f7v;U"/)J%,& Ñ/7?9] Kk4~;-m~"pm#9(ɏ .ʸRh^~J`fwSh˞iRv.{fMNx\%v *!PLEid,UibJ)gؗU4YGP) -*Id,ڬejVi%ʨ5}=MW:jJ )?LOwɡz^I*ۀZCuZm"aώ,޵<߈YO.IĒ\H^yst`^-?֟-"첷L,2)޳MOA]$M&ʶ~lc!k^\<ݣVIԹv䔪Ty Ν>8tջ/~?(eJK%!iu3*iIG%#(/mTEYڒxy6Lh#Ca,~T瓏rr,њ|\z qAy qKғZ h)zgGq_ %D*F3;©烢[l8 2N.~l\٬9Ωke$f9YJJgKiB'}^a@;Fq2ڱ%[60c$YK)679v B1x߄\ tx0h*x] aEv=rZ@-uR6zBwɠJH:-eX U7`%Z54T[/0`^7lŬW&4FT~zh&eN< bR.UNPV 愈Q5ֈ*J*IqjKj. a2c&RM1:#ªŎ;ΠК&}TZxUhĢ3AP(5F*"52T;UI+Bۻ6a , df^Wrdɑڏb#%?ڲb-[af)]M~dXd΀~5RjDB*N=r5 fd|x5t+fIhoF%:攞60e.X! .QLT`A9+ځq̫\Van#w=_WH1E{b|+IpMgr>Q8br@i^/~=9+q$ >|g!hK 7ق@[=㻓w/*ä Ba ӻ__0)J5| !|$0dt b8֌Z/ %E" 慧i-fƺb/.cfc 939E(̭qe@aL t{+r^9s$y.KjȖFm>/Bْt#KwBohZ }Ǧ8X6hҷsG}UŠ— hJ 6O fIj"~g fOe:Ͽ]Oұ sDT:LUGE>V0]P#c>2pƭp8Ug_ 68ݴ&2BfϞLǤ@jI$'^/#9vDlIWtʹSc@yv:)Osp՛#fmCKVYI%52j-eİR*wR3 _ D7I0`d).q/ֲF+l ʴ8gHp"@d@Zֈ9k~(+]Jt VV@JɷkL*hmu~O#ȼ )#ϗD'V ;5>4{siW-Hw=(ЈM.*KPJV2$ dS^(iE1s\e.MHԒ.  ߲.bPNq09WVWkE.}ET+eK1o˹6e"Uj;Tʖk砢DtA+nz [ybغh ;LMɷY[\N? _t h sxaʭ`JPYoQͭئnTzd_Z%Zc!p й,ܹ`٧,|OM8Zi^+{a`pXրD:D4puY2L0 nf&H|3YߦoS6"HxjdT= [tn[{fa y(V`>0Eէ6HV52k{1Μ IK-l0}9/8k/STiԪ {w%u[5Iû`Pt7_G8h 9{љ ztPŸ +5)AW$jdpT)$Qbc$6yL}-C)浧)ӄu,oX $ϵ'Luݕ-aJ9[uDk"_اyu@KŇ<̠Hf9N5D #w2@l@Yס1G/g7O:uUvӓf7}ئǠxx2Y>\^M&`jRM~ZGGJ w=hiT1BPuNrm0n4`^,u1 R9Eض R̒;j7MS9 k:8X&c?-OFni4 168Uh<]2,L˜Fbq>53`/[%WT$y7nϱ62-c+΂ iA i$ &LϟHl,Q;"GggKm9E!6nߦo8TK=kџG*6h\j{DJ݁a>H N`J\;h8m U}{Q`9*%3p@i $\Kl-kϮD.jl?ugnP6E>@DX 6Hű^vLyR`,q#AdUcWV[,Á)%CDLL (;I7ձ O}mŀC j٣ `M ZX/b}t< :Tubڡ-h40dl\ xK-dHJv|;bGs)ǁMfwPcvHWB\MuII?{dRv;0v"LN7ɞ,ewe<.uG@H@^I&8>wHkR:Zh'̖;K5CqRGp&ue Q# .4F3 F}A"S蔺? T``L)_hkd>$(}?P0I[4*2TWagPRJBu7\7+`C(h#8+s@EsfHgވlם<>j)D=FW {Tz8h2ajDd8*kg_v9Z\YrNw"1|+4)L B rAJe$'\Gz=81YOq lL\VO}K󨴁`rgDxoD^֎<H)!P@(,pH:ols . G^m@2 qGd䆜.L%c:un+ ̘${Үq&>Ć˦n>;ܤ33k3cHnW L $8s.JTaφ`AuDDb)D-# WS}uB re~a ` 8 EU# lϛy>w4F3:4 6T9aCRS(hˡCk\PI 4Ls{:kX9utYq}Fٗ24LokD2ռ~ w8L}KuTO~ 0.ut:xO͓3x5s]Ta0bdϫI*[GGX6[;Ŷ+ri*gaz]n>,^s1v0Ȕ}>\8>Xz#ᝧ ; {a;ɀz>=WIFYQ >mм,)FgՆ\mҙׂǬ!+< h 0,nlbLuN` ln*ZEJ/v;Z!8p@H檬hwo&pN!l36hL:zjd'M}u^2 nF =ufv)Nӑ V-G:FF~E(*tjWic^#\t0ѡSv)JFwmz9gkUр$HrѼٞm,y&-ʲպ9[r<6e$nHV}U˫c'Aiu,3]й=ԱM#3̎aN#ǠO|uK}f$-LzyXu5-CS8! jZ)sΗ7v=aC{_ t[?}꤬h夼FXt;~7_NmV<5i56L ӧV<278xjړA<Ɂb@S~7Kd]WY5s}̪5_0}7?Fvo R)m-H9R=+u]eUX\ﯮwMHGu5 %Ӂ`䛓෹}!uf3t@xuz y:W/ϯ8lYmF͗'P]؟J_ >^>>`M ;w|S7=ZVwp/}b!wwg'Mvg?ѿp$)6 SfA,-5dl! u^]Z n~|v(H-Ũm&0^ՠt$O(LN6%Vض.Ճ"_=) $uH·Ƞ+b ez&&%E@d͹:1` Ԅ[S8Ec ~w  KiRիaVyi/j0$G Wհ>7w0ڶH1pvi܀ݓUM8MXii[+Hߞp}֧S>p F u;mSV X9I{l0sΏ>7extӆil~c8Ӥ=sZ y}S !_a325ta|ñf??H*5;;ui2۲/g,Tڋkn5ˎК)f~v3Moy lƯ=!ǯoggs5Oܾtpo%A+DT'"B9܎kKd*K̲g q޼Ɉ9zր Nڂٞ!aIu Fs|AYy蛏?awWT6f5h}T0[vPP}g ӂnc|V Zh|8e5+!;d5{hz#t$1gxS7%|F24=4 va\MpݒxAǐMnt0Jߕ=%5R\a()Be60\rO T%DƤXyɒ՜^6 ikCUm<e*qٲ.jc:4nyx͏[j;w=# ՒB,<kZ9rR0ZӪZ%yY0mL/[K#z OMx~Iɫ1YIR*kk2R$Ԅ7%䢻=Xㅣ@WX^rf+5<@φ{| )~#;f& PEU|Fv|ܗ判2ɠQ9&5 1Y[RJoBYmTB?O:]D  (y+4* , H4=ϤͰ.Bo}ŅzlF7tWZq.A+-:M^IE2 .9l%ūfa!=k"/1fNVV+W.f9͈G,qPGa[VGVvV˞#?hYmBl2Z 3 ypdx-L"qQ٘#HKF`16 2)ٝ5衕r{|2X+])?}RBȭ$=VBL ˜ @Õ:.&I+[V5`tEݳsOt2 e"&QJ2NȺUKL E$ˍUX3qQՊ穄jFf\~:)g o˜FJЭ$5^*>I, ds'M}Z +V]=4y5ja%1:&dc7æ\rO(ݑz`-kՐ2Ie 4CcM[UMdvZeea a\)裵 JDJC$Z"%bVdRA+~5f|y}~"%#ENJN#H$ 9::44*0w-h6‹`},HwbVYZ)zѐrKih!3e6 0HCf-ñW)}zϩ֚}K¸欠qtŋ{H7&XJ ېu=rO"oovt4cNGÃ)}/!R9EY{7 {WK.S=|(8&!/;\8z:1)(ihNO@YG 0h}4s4Oix{`Sd[|Po}mRV|)m17 !HԍS:5>F$c(s39[ j٣ԭY}ٸ\옝mB᧏D*.`c d-tM,ZM"LZ`qh l *iҬ Y0t"Èz^ BP seEV](xB3:}8] KKūVkq8of+6d\tAFՀJnptEg7}yd0nJyii= ,&) AFO.QpHL҃"L<$,D$Cĵ f)%5YՍzgE6+)[sl)!ϯ) " *Uĺ@':RURMfOSewh 1\:.OdN_MdٰH/H/]r<ȬonFwM#>n7HHMJ^T1_'oyj¸}f,Hn8^L?VV}-sƫm1WqަXM̘bQlv#3֥I=KZQR@A2S$-+YT>;:|>9ڗPC+{9&sٽdWJy.gP(dȻ崆EgjihMF ٹȎ}8F-yF=rdVOw*YeG ^3x{YTs4oW?9kKC'kX{~q]dv/'1ʷ'y5{)y~Hl]ut'' p;WJ'")Izެ F z1crs&F wY1F৳7e Xk _1whC&[ƹP*XW D%T*B ЃŨ(¨(˜lqa+w@1S6J܌IH (lon;.=VME~;ED̯أw Vahz; {Vŋ'-~8>%./zQK.C"{ɭ";`ÇżsG)L KS@kFҐb2ݸy,.%v z֊Lz/"9'/))UoW~|#5䷆>wMf4eFp7ײ;Z44wkW]v,Eֵ!FϚ8QoO'<8w̭sn*~<_b^Va>?/y) ȴ.Yݕ`'= zr:&N=nmS.J" aA t9b\x(fɍ^θh`j।9WQ-Kk>`!mCVmҼmPƦkXd؈[]AGK2GtW<( @4cvڐi< yphC?Ѩ!Hduo=:WZ85yh/tf,' O[Fň3-4`+cCVjTk >C IlaŒ5lS ֔(&-)v%o7,) "UߺuIrw[QlY"Ki9_,NF*i'TƘi_!{ujգkU.رݐrɩ9w6b\o%m{ZQ_9SE.Zpswo4z_ؑF%2+$Q+%ro՘ JaȦB[־'ddn>v qNn}5AH˿t9 >ftho9zw޺DlFu8s@;aa2cGTq$_LtiՂяzpl_[Y>q166p'0GuKIc<^{p~nx~\(2>B\FްrwY 1]KX[+!680vy;C@Tk 8콑RW6.7oO?Yu Ҫl/ʒu5Iaݗ<m ` ;E`}e1?͓)Œ~4?ؼĄ^"̪sV9#Ҧ}FQ RX* -_y-*¼@ѻcfܗ>RM FsPTN3):&Q' ETBrg!쵝Z̩.BaeмÚ;5T~HS鮷Z7O,Kžfۻm!@V.dlY+DAL`sҨfS_""x<6̙JI f?~inj78U6QnڈTjp).R!l= JW}li3NX ?G]TѸaё/dW;F*|︼55l`&HCW@$9oe (޸<7?*^٢בhRp!T7-ꔝ0^]⺘s_bYt]H9ڈ"g dEJ!TÑڐ3؂ v|`ZUyOvs< UytyG\NC?ѕ!ɣ/G!O3}s-oB?GU((ٛ//{sFz{[oݠ_l{0` Bd5elD~!Prikm\"-L~y//h=u_rr<3w_?:;Øۍ2npmf68~!X`s _Fb=gì ̈́l{C!o=`j#~:iMGŴb/yBn"{Ik@ i|u TMS' K]L-!cX\f?$B9j,q4BG>nl/ Y<Lřr[pxpڐ3"n IO$i}|}GӏC\|zuš-zcKEe,fseʛ$Dm 7*-0h!Q3/9FC$ۼq)9@NG^4@L4OЙ8CGOOU8 XZC ^Rh5 eBk5z)krm){KٓPwS]]lMZQ/3F~tH38uy/U+Jc'tm\!WDEFY; h3EȨdQk&gq~VJ5!{2BKR9cfK>z:v0ފkV5oC]vO0g4ŦVKmSN9Nl-ۥֽ,: 5rndhD..K'E2ޅ!V -i0W;,AAX0֠_~w8-o: 1`<2/kr Z|XF>3wlօ |3L杻_1̌]/?{̟xX<~8a޺P}Y8Pr9h۪k#Id&!Mј*P( pg 8w$#5ǘ Rk`lPVB5-&v&Ix%/`JW?`FYQGBKMXEe.1wpڎ3gpArda\7Hm*jol#3vݶ e͵{ճ_CO%  y*grԸ(kb A[P [WjmR`OP y{mj'\pvb48:z=&n?Hz_ẅ́VI;nxΦ/cQ,ݿpтd<#o_$fcĩEU+=gσmG'U=<:x?zw&UfPv}X*f 5ᛗGl:b>Ě԰ʣϴ~J`XFq&!5avF1}(j'Ƒ#٠ޫGZ#+JR'B"Q]Q \jT4RY|c9J}1Nޜ햂 q_Op݃) ̍:Q9h \sFzH&}-I]}'/1,f)CM L¹u V9E5}F36`!B&8Rgh9,1)hܪٳ oS0/GWm,F^416̄YER|Z{:TD,vEB.L1'JSgH}qvNbHzæ/lL>%;&_*O8*}|9&}d+HQpYMs#PROս)5׭ka*Iw-wE86KÞ3]Q]@D(*rYdvj*\PHW*wO"wi߷OEvvMX9LX"$[ S1FT(n~E{bLJ'"g-νMPvL4Y9\#1TJM 0TC WE]ȉTPCjT~O5|r\x$,nԭE7RE܌nu4Z+񜱜3ۘ=Gtn$^@D׏[M 9{CO+lj>,n_>qn }QpݙHF>-Z)©c\JJ -,`fZrAcII<|9n?)$sUNͿ 22sM>׼H(TbNpy089/m#Mg9{{]{Q f,F2XAFK+w_J(_z=P13Tm֦PA TzWK͇> ͱŪ{P7 5K(AհrOu<4VcjNq/٤4] 4и**{gG sqVZ)ǞUe˥7v]sW}NFK|+G-[]L=}/. z51=.A-L)N_>P*\ 3e*#FdʑnU.@1id~Nyj8Y 6UR=QJU"M"ꞃ7x`GqP"ӖzU$>9xC)y੒ouorXtE+ ^So3aWZݘLYL\2h{ҵB}Z*TO ׷xE[C˧]?O 2TPM#J|n$lLIV1xj!A98!L%'C{e @nQDFgG6JMbT"$4)}lHѫ:ZI!wlK hDa+I13\326<\N!QKY/և4(]H&EA%`(Ua>A̵[=:Ι xrif m{bǬ4\&1?Qܓy{No? 'y}x%r\-n<@rȱ*u!s{ {w[A;#2h%U֠@ǽHNk }墟E XRեSSvuEҭĄVBX.O*tt^/״;*hWGoW1Ů=J#q=>d]ʴo ٻU+|)*=\rulULbM`kcdHʖS+aM"+0R6F JJ o(zYRZ$Sqв^9yGE&I ?[IE0[!DIˤ>C(ީNvx >~Jk붶ͅtE}8O٫=e^w}"L`PtsדV}7J_>y9֨' J XQ\ }viTB}p_6ϵy_[Ibd[QʹN/4(h}WՅO¨7:|8~/Z܊᥉^un—?FkixLuM[(e{j)!="cFOF?NVU@PjO,x89ӵRy0hi=;dQE+i"C=02 bObHo^䙄G4D0pL"n%HrS!6G裪BZ>""9{#WՕ[ܲ1/GX6o/X"ϧ'˫"_f[Xk9f5]8igu O҅P]3׿tM /#| rFzɾ]}[;]HN},f`nFp%n髻gYV;>m`zB򀅅"H6z[\ EUQ%Ez->X!nɸ$M;E}n$nY%Fѕuqt!p4]r^%42c.^ԱԹŭ1JXȋ;oT?q]윝y0wB'L8o5cP_mK]05l"w%{J  0g˿[/!yدbF9l7;)s"v~Vݗyٍ{ LP׺a~XX?[tpf`ėoG]1[ i5HoNՇl ؍윛av?FUa 7bhC;e#A|9^Y]#-Nwnvw輝dav_9ǣۘKv‘1n~c&vv?9c%E芃|fd[Vfd |fyMy6nj9 ~۝ 충7Aӱ?;S6 Kvčv[uAfd7#ݪOdN(NyRܚCgqK՜h X*ȗ`O,ҿƽ~nsʻ׋>*JBGfI-l(ώ>4L+jSڗ剖J׿ʉAᕊΧxwϛJ zC)H2zol=p "sh0Lqp;`7*ĽNJzR~uͼPDE `X{cRj#ز3W﯍AF9*$HO0/6k:nA7bу,<?W⛐ ֍HRZ#_Ġh&7-+1_lb0wKT;V(O>O?G;JIxpzu^pYuVj]SY{pSetU_M&}%!?{WHr/]yꅬ!qHCt#ɞ XlR[ebVd|HWO:=sك.PkOy@U"g YuZ!4 &Ck. +KyCO 41" ʇɻ$. "鴢f˷-ζl*YA[ಅB Rpl͖ly~#,z:iuu62,X|g&+`*ֺc&BldbJd9M8p@XHdyTH}s+!4)^ ِWK .\<#Ju!Nِ;b="S(+p@PCnElv1a\)p8*4n bsBvJi bW1ߣ[G׉IKRːC2J+!x3]ZE9DPz'#U$CR9)Q]<>e PWZ#UψOI-0>;ċ;mN\ IK|jmufMI+C&^Af iz2;quEȕ(8ɥh$C vqUV'ćB;gs =^دwz )A%+Y~dJJɓV4WL $m !46(&=oHr>r =6pOX8FH%^-iay4hٷ+m.@.06yge|dx=Y kc N(=lRkP& !٘"]A -:_ F "_O@+U'[yff:69WGK%i̥%h C3d$G H$ hҽ%ZfB AwzW<;|>FδjYU{UgUA2Yv)sVY;:.w 8ݝwdU~[ͤMcl!)bar`8#j+RR]?~@8Ћ6_Φ<|Wg, e8?xY˭Ķ~| :G)g+Mbr]81}#4<',}Yk;zxYtBȟ'izhOqXhA?n.y; "P&I j2;BvܳL%?y9݊3%}Ç?t:JsxmlZb[;e7TGBM!~1oezdȹogG|8!W+4Z:rgw!jI,ApFbLsY5R}oJsX}۲[YRN^> U3'{lod9c锣8M;ۯ16T)cԊg$N%3b+X:6$1vck^!d6!/Ȇ#9 vg \kNizPӀ%X\jc{q%v@⁊zĖxP=aQAn:[qک+*\1 ˉ``]8mS`Za=q;Zc mc뗃uxk4>8Rx PZS !'BWRt(ZϬϑ~΢~]*vޏ#&-;(ŜwY%J~|z1j3(As8NhTZ8O)%&ikzHXYസ5lN.Bt˝Y#?O`*S,h2"Gi~U^qTJ851ܹ2EA qP99 r=*Qob™rc6 RrNCj~cٞ.-k5ȫp(ZꚶFvLxNka'Ѷf+CTA.2Lg[35Bʦp6v i~x4 ;<Zn9Rx"5GUJ)$R!c2ZfJ9CY(+ڙnɖyF1"_.#(_ o]ƽw/,A!(Yc\d{Hbi,2TbY5Z9AUE{IMOo_Vq~*]ۅk6 ?i ҍGtmeEk~_+VJef1D&8J5LuFbՕ? ~s~l/5Rœ%vݱwґ,[p UP} w7-Wx.(ի+ߐCL?[7f/e'K< ݹOQrӲG30QeyPoא>( N=g>UUw?׿ϛSKKrir9J濓2tv,z?O;px|9F$ 3ޱ,QқdD7v> Rceʣp6c AZ 4jd }SX6=\GU`!k]wvmYVruFy6:>+u;G͒x--)QԱ!D^˴b % nlÒ44?㟿wB^q"O>\q([5/:;'Мo}-'6xr5Np=>lwFOgg߹gS .ۜ)Qqu&ZlyCȎsn!8t!HLl\`-B9l-ɿs>!+[IӭwmMy5Vp[ͫD4SO>^ie eH@MѬHxZ:!m2nA[דkMh'pip pP:rVX\sMZ:]ٶyO^@xZ[YE[A~+U[*WY9 rsU$j,'vz. yWVxk6_~9VW$\o2\+pJr =,|I JXPH yH$(Ή4̦&jOeޝi{Np;wPh,b\ܗEd̖F?!GI:&]ގmly,Ct:{R:7$&CIގk^P2vIB5vG-1DD鈈s^reh!׫1=Z `ikmX9g3T߻ڀě b,H`eڦM E8o5IICr(pH [/ȹtWUWW}]2* D d>R `% q& u=SSkXVKh.h ip"X/CU4BlΙ:M(%B)[4ЗhR&Ɠ̒F9Zc((}( Ri"qkȲzµ&X0sVbNhgrc7k5Zmkk:\A[tkĚsEh=dd /{(yT  g$g+¼O@br@xJop_Z tM ܐ#ewoM$jJ|t]M6þ,SC$@x8d<4)2F$*%WVװwɐJDܾLRTI I9 77k^q]{#YL2ݦz e)*ŝk4lZ~J-J!nP&"R##JmS7g7PP'18`8ׇ3ܒ3roɸcKR܋w h0n?U*5˦,ТsyoDyn⼡gt]xPI@Sl<6!?3MMj{VIg,975ILi" )$=?iN:H&3Gq^FHfld`/H]$-wzl47F2R쾅w^"O8Ȗ_hȥ#§eZL)q\+=A|qSD~_Ƈ}, J\Z;مT Q(ut~~}N BzSDau&:i^\zsbJ [# VPH,$c~2袠#ZQQk5⃻;MAy\pOR\3O"5U7eH L@Q PP~(:"CA'!"Q` dΣLPP֫>-y|6״RH,9 ^}ʈ49x1vV.2G; j>^}N/(($T(JCR:OI!"„ f-T(݂`"HWJ%<). 0oJ&@J JeJ%Ћ^y0WKҁ86BaN2>Ig}C^y/[|2~X35 9'I׳}BF:_0D…(9/7YBF~{x)t瓕5u";|u^OQ֦if]]b-޽\Ï_Yui!,$3 նqYE6Єל(#!gMJRKw TzJy\,2 ;Ӂ uѼ.  H ps1P*%8;yi4w2Z$-E%J5E0gXN4/@j }5%IjDM35;@=pm" E2%:؆K6T5D6Y"Si^-,桛вZl2K?nm${nhM|%j(7[;ˋ5NtaF7cue(l|ڛ3~0&}?t53-p*0Y}b2wga`~9Vƌ2'0~a Hι sE_G(*\ęFI”Ϲڻ|6톪`U?U{+ѴW(Q:wjDm:EnTNQ@;EmG;#(<[lMТ,&nXNڕvժv5OjZyw^&Τ]Hls<<36c|p08/ b}2ɤ(lp %X@*kޱTrKք,`& kĢMETO9c: mFZ_"ի%쬲/Jp Zjh>>JT~-AzFŸ"E۰CRb6){Y!Y&,r&b'n5;GOy@ʼn['6js֍X\WV` 928cCrhP:#G9ZĤZbJO_a<y6=#w?\F>it Jks,1hiyߦܐ@*O/In6e.f`2E l}'e2~WߴQ(=8~d|53f<_Fh}'T0ZdHL;0FUBѳi7UsBtj)8c&KȦ]~CJi߷N5MEuH+wLRg=`&+\&;4TK_rCr9{gaO[7Ritnm:b9@YS,ƭ16P[e]ͥNSULId&6J"(cPl s@tH0-ʭuܰȢY"KSc;i٨bKMn=%qqLa/RJUzԣIaJsFۦCLw!JxUr3]gy|rmTIGfծMhd|6-`f_}94;Ng7:O+^y\1LN+ξSOp%Y+_.hUi ?.ݯW^S[W ZRN|h{ tG?ن\&;|AZvzd5.+nQE޽q0t-ii\ .|Zzdq?%. i-)R,T /ȍFx!wi]GC|.`C?qa0ŭ V0y N܍O=~a;x;h,<8ݫc.g.uoқj!n?;p+ҕНw ]) ؂Ti S^`I5/{n` Ad])%U1ܨijq\C=^yzZ)/vś|f!@Jxsxq}7~uu8ƣJZ5xog |3c2}8-yRIfFߌSw.hluM-MLK_i^hI*cQZ* t[NeT 욑#K@U ) T6nfo?@^ߔ~B}}<#f"^|՛=A5۳^0a/cn{ܖؘ \9AZ~w߳_uSL䶜{?fj~ebyQC2bR&$11Uֳ _xI:v)G:M, s!U*m W,uât|^<ߺfqM{?VR'g-}(i$yVO oÓgB?y$Lo0QU7l)o-7,~QަroS9j9x0*pXNE&X4*ALFm\i /htYlԍcXNzC;|;$uY[$pf83uKSάI3d)0][nф"bE}r0® ({⊷efIba/oʚ*wDy": !ŖCZc' N-܆8 )l SKE2AvZ9OD @»L3Yn>\1% Ö hB1Q5Tb:UZ|*c:`'BZ=U/hi<%V"p '˲D%)139'"3΂erTH )uif3Ufʔu&}mAVҳ2(ӳLAbX@/B~XƘⰈa`9 sHM9ZʙQrvVk0sFJ .9yRdP! uhciie ?Q+j `d;\i'0_(P~mw9(b *cI=OSJ )r `MRd0%U^6b <^I%`àraS+҂b81D)&!YT&))X)*V7BT~bGX 19o<{:kpZmyBy ͣӕXqlS|2]&^`=:_qR;A[7lg8UiՆ۔Hc3$ p~Ǖg 3[z*lqmk*ON;~QneX'O0p#9K 檁 mѸ~,S)URyNv1 7gHŜLJ Q Q9&BP4 (>d`PP*\ kIB z4{?`"c*}5CbP->XPf\ǁqþK|.ir%UZ;k'󫛀Ff@=?qPXvPbl{b{2F'm7y`YFM9Ei e[l" C$*Q15(swXoԇxo7F[aNؖB$.!BꄉL'|BHéB(4Fʳy(`!sMƽ,O27˼BaEe≡ iibS-+Mxևj5[d_7}iA "htv6?Rxݜ``a0MٷsZqS^LyʔY(1 A#* ybij-23L ,3 l}z:a\-e2FJŔj 2f/c0a̬)>WGlmvOuLo#FBஒZ \j&X* GŖ%1Aф$ +Z*eb'*ư ٲ\z7[%2-Hd]"C 5 i5 G gY(ބRu !aن3VN #!\"[}WhBwj:x/}'~o3Ln`sn6 idخM{\^MM/'Ri/Җ'V"kx0ܡ\f`EElPM62$[F:P,ա<~!{$`>WU F7`hj&.Dr%`v݄ZХ[y]xOxx}2Luק秽 8 7bҙLh,/i''73Aqa<a:xҦ׺ڍmnY)$C'*O  eAgzJ,NS-@-ιfN̓nvUJb^єY[Y6L }C.Ip 5H+?΂(5mcIC%RvfBd`Hn `Z:bWW{߼lYo(ʚOpOobes ' (L$5s8faXE<7Flu4HQ֌>n)>i@8p#9zF~9%"i M΅Q-%;#y^ $<WE(mQZ8 6&gFF3RSað끬09a, ~&d rc< -GQVfYwd.Yg=>=x" sήO_+zzt^|_t7 7'ПsX…/_Ϟ n*W`0>r]ʿ<]&ߞxMHw?uxzq~zVjT~~.H7G,d}1zY&'.p&-.MGM^볏#8$RLq3~ܡNZ+L7KU45?_x-<*ܱ~=-'zS\W]\z>.׽AߠβݞO%bre' }OoR8ɀݟV.|ybF&2J/-LgOaݿ?FrܻQ|l9g{/Fy9@3pLzyWsAB/!3W=3/i79) Bi!  6eCXb4z<׾E,\|~w D/ |J!>ɚ0P$///BLO(/7ELc#+zXibdSwZQqF{yQuSq"%y"B#&bd;jp**U=b G\ [ 8/^^ʬNu7: ose,7+k1\/m|Ow xgưKhY 4v"h;),IGMI&!_5ɣ&y$7jwz?I*IAplHyJAA`hť?*)1JQ%72Bђ#z%Hj5=@QB,Nu81)nŒ2GΥ(jaLզڤ{ks_w\B g~R/BjEHMW+,2P^2i3!@,0(6YKV;e9τkĽ_oV۾1j$9ٚGV2Cr=R1I1 [>S+&cl2"R! <rHEb.V Z~ Q0i^V4Xg-xضvTLD 2hmAf?{Wɍ/lxh_ vvԆֺ%5txm/o0$e]Uݽd| I%j!eɑ0 ԐNkITv q^eo!QW=#o.@ؘ)ßRqF*^*U f %naϖ*+fQ8StrBD3O.GyL$Of&i%g2 <}i Il$Qi4fFUؽDf4[hdƃ)*5)rjTz݂"0KD2UE3áv"|pqH!1hL}R0$K*z+T!vFrbN[crT)XǘW[2:Uؽ4l,a*8o:3!xoM 4@GmԒ0hrˌ23K"شd#DD` fҽrm.b'F ]8EV% 3F28SF-ɹ8sQ+t.^Ks4z!qDq;\]eW|;#rkΛ}3d$pC`Gqݒ~,į?y_w3cl}I43S pxlV+q 6_ݤ]\ÈˏԌ =cb\ur8tw_W.nn67 ͺ3^w_|pwY.𵟹^/rP8Z?}͟9+izت}W$7 d-V5bqu^|ݮ2k:ֈ8OUeQf1D cb ! _/< @)6Rpj+b yG3tT d`F^V:U4;AH:>cOqgoEȤՇ\W_ۖTK+g/iCtWSѭ|M Gf1fj*z0nyfUxe:mc٧YR|7ߎ[ ,f@Osh`~E*ivnzAI}H8"XWxɇbAh|K5W'Z.-0Km 9"Jv2Rp+hοKڞzq4wwf|La18[[i-k߮߾7z_teӘz)7%iD:ħHh,g?(!"oj}vG!qQLNr"OT49we}1~/LJ|ݴ8.?Ϟ ycnċUXEƱU?}tAZ#VMcA'"-e>]ܮGJ"(h0/PY=1M.i "DO37l=kunX3wL`Yo96A{ck-X͙ș_|\ \Q`sOι\5aLDfz.4~ћqhgL%Pens)j26!(j I!n֐Fs5*:CXd::8\sҫ>ɁG%*e $OwsC+i#BA65Pv9idCsc] V^I4PA즎aV# j#3 50ɸ&HH5L 'l=3A rxxNv W`k$B=4KZ3Oqm%Z`7d#uy+b1O1X<=\7yn&#vyjvsmv,ӲFHꡒ˫=;ӯ>_Η_- > R9d4";,<ΣGFuL"8mj )kGq @*#Mnlr {T YdEߙr.4CZd mezP04@UJ沜²j{W1[#BX=]m; P@m8l \B)51ℭK">}-};tF?!|E9Ç8k/\̮_7-rv:s6R=/?JHԒ?_!ӜegK""c`k9 _IiyͬmG_8Z YbC8|m 䣸cuQS~zʽqE~-|ᬗ#BΗ qH[H EVu2Zf$&ƞ 7qvauǡ&gkԻV1`v*$cI$tpL`$I 餂sygJI#=6\6]^{=Dj5Y 7m~mC6mݴl ezu&JMXϔAEfXTT(P2|Pvp𜂤WSFcY(9*)B|vz-L(!E/Y ߜs^+c4_|{lဘv@&:(ƾl7{&: 8MxMvO#Q$K6"<-.,Dh§&:0 N6}k]t/%WbiՖ뻿tk`kr"e^[za0#E;..طw'oИj_َ_VaUy: 4lSuǍ#1 hzOnڭ괾cv;A]dڭT@v!_8E8xv3 OBbPCw n" 1vkV/{s0ujB_o?_Y2d;AXbjyv|U׼%*`\L&̇|U+Xw4N) -a'm"esa՝$/Ҫb z㷿6[|wc?Ćyb`%t\q?]|@lx%xvd!}WwsŸ/_}^,åaOL4OG#hxZ;,dxR^ c[kA0 $/XOۃv#@J"?*fOa9uIA9A`JB(0eSJvNdcͪ^A' C޽UXUVsU@ndJ7YJ3tɤHڣ-IG!P]rUнf0/#/ʢ櫡HH8ȋH>Iwy/ׂsI 44`M:+y M=2JEAT;ZNG3Κ6GoI,ݒ=}xV j^Z1$0xfS& jN !_8E[8̞ϰ8#jV j!x~`&dڭYZ6Cppj;E8s ڭ괾cv;Fc*z{n͂j1StSSv$ ڐ],B'K\{6\v1(t 3mI9Hztbcg>ų G# Lcr< [j$FJĬ GKr-%}Ji&>@Q818jT^D+&haBor4\ c Pp@Jcg6 ps2#Wr3'W ~Z"A9 ?8O~-ּJAp9*n'K!jfl;X<ŪFaSJVa V3,$SBۻ1R!# {i̿TdӍR ?!ɥÒK =5y>}CG6}0psQ.O;~箋 ߯Vb.VXw0L+I6|"|HƎBUAF)m9lh$G@k3 Fh ! hҒbn<-= 8߳ms{]HtWrv2#.Xg1)w*&iDĊ,5(eC%DJ "ykx/bPsòC.qq R?`N&NYyv$SbPCw nGJCkdm |mR'oqnh7z8Z1N;h#z4z͂j1StoNiV8U.EvTtn;Qs.Ls|,:RW_~Ba{+wn?JˇXg5ECkM:x_T_|w9fڒ`2<}8 V[tv7˛opvg:]GvW}L bލF%0:%s0$GkNQ}K@ȜfW{WUzu&k\f/Fy[dqeKД.R:3eJyN7?NǰxYy4I*jNB|S3dP>6#zXA5%HiD@PXQkX":Jb%y_#C~3n6 |<0gJ͗n-5?lԈR#.LjkoDP5QXJZXbz<|a}it,xA}v|\*+V{|' :]XLWC`ԭawQp,m)_fy ƧO]K*⸣FTd(v7>d.̗]wQcR,QǡF#Y=evM.O|X!9jT:-zQFyٝ6~Ӭ1b兾1 1kgVHn|Ln3M>3)]}~͡F"м0i8Ԩ1 `%HF ֽ4JO/XUb1Fr)Fj S_]j7Vz]g5_\s8~w:p~y jeΛ. bD@>ZkSŻjycUcK;/{ 1׍ aVJu 1`f TUT}z\6oƴ3g[77֪kW Ӽ"kw*Y[=;ZD9Ò|@dDZgv}8=΄8 d`PH3!G$+I։67mcB#-6KMT/_}ej.=C .84zb_'w*ҋHS;#]܀IrntN%:Q%vw?͞S;j/c$Kw 5whz3NYY0USna80DRoܢPsNHN&vD2I򯴯pr|M(y"U+qXTs )vbi0QFӈ4Öj鵳&D-(D[2ѓ5Դk.g"]9*]&.'\vZqɖa9xu9&ՑLH3Ό.F) Z5LQ,21`Ac'UQ@`i!`E$%/ rP;߯BEaQDn|M糏ſs+Ho9߷xꓙ_LWUj:u83dbD|7"W$_rhG_i/Z,`DvڭgǶlVw,j/-i)zwLotB|dBK J9aFN*!L%B$j%-1Aպj]{˵߽\[u]fdoh?a}``kd\$To$^-"z̯pok7N>iԪyӻ&~-"K!$"QD(ֺQ2/)Z]Cl-| vڭ^}J#,z'b,Z%Iv@Pmn3 Rԡ=Ӛ$%> { )4H aU&!Z:gK0-cEY[Pz'b,|x1J kKRRJQWMG})b+O)wzLI-z )EhI$쁄#14Dp17\I!)jhB"h$D!C$DI d08 t&?\FrgP GsoMࡊ( RD˷fEp#Ei=ߚb''UN@/7DwNS#@Uj1/ϢBٻc@Tn FESf"Ma=XD!bP@ov `],wj+!y`ŏʁN溸.|~0rM˒WBo,i"F9FuUU&L1@]vyKt׵OPNE׃to׃Eh(5Z&cZ}kjE}kN'F&tj&NPEYOfI@*B.eDk^H8V 1ty+{<$o=&&yr8U;)0#):uT7;ByS(%yyp޲NĨ?ӻft2xoy ϟӢ'hiw_(к|?8;`HU'`DJ#3!+ :K58%h N:w)_99jZ`"#]Hm-־Y[x2\J["ԸD{kzJǴSyNIG;.PLjA $g [0r<ŪϜkN׮7{7W=߭73kƄy-itm$e,"AQ+P(r& ʢ['YoP]haGLªf?̲) B&tӍJ'=QF`e>jsspM~];B9fG`qaLGfTXT`#)E ~<259:^: jY!j m4`o _c }:P| ,n8$Q1EX9( 㤷>`g[E*"DW[cE'/Ñ3E="9(8gdw>Y{O[HسcVYfQ6_{HHZf 1k%~|X"0&2HPc[E_-f-'4X7*'%9@+O!X܌nQ#*f8%Pu$pDhWn~Uge}>.N8;zp>*r'סּ ˎ~v 1ϧXogXj.'&.| a6/&a>-H `q-+Sx-i O^ hODn13., Odf.1"^*R: nhL4>,ꟾ6/CXߏ3+GZ~294՜hYy]>Dfowӛǿ|TSi^~b;~_y\'"cwyEOGƿ;qwΩ\5○/m/} ⇰Xz]sd29󹨶;W s7y:c.u;"E RrHd 3"˝VcBNUAv%L)r'35lzu!Yv}~LOЊ_MSningDdd$yg}B%!f"$y{k|W}\~Q/ޅ4,A/)p+~k!}X8 $%JԷÛviSB;<1P"VY-!'Gc:rx`X[w aBZ; $ЌcbR#c"V10C( 0F+v6<,)0.0fz >S,^2ARW&l).JU6 0.qȷ┥REBgb)pbZa˱Dͽ ͙%+tT^[~iME"sWMZ?xeaK +nbrZm1k"=33"J̨m)3ߕۭ#,xiSqD?_nOd~mBW\m5]5)"iF5=J+EEcE@G=־$bиC߾mEi'Gۗ#}^i5'  QqX1S s(vhA4'% ,Tl~fݠԔ 4R$Wahh1>XP?|->EkwzQn{hD<|,Jg /j,/~-!O?Dk)WoE=b@1{-[~qD.z疣?Y_yV(6~.W4j19>ᑩt{H@&Cu(v+*:p@kc$x獶<$FaOGy:40G17)|;2#KEC(HX _z{tdūٗM ji+ͭIٲך9]nf{Vՙ7zoǫ4|g^+-8#>50qc!YeӺ>d_~vM]-MnטdFN2<2bK𣏕@\isЌ^hTUyXpqtyKZuK!/Η'ȭz̍ ` Lh~^B3"S]#q=+&$'KD d#qB T)fנ-q*ՅG٫ m9!t ntLs}N0òdU*!G f|{iF*F笱&PLXoZH ]E_-e P^RGٕ^n"$ u슖eygnڒc:KvLOO(^A^^ F3w}WQ&[7op)9ji0e4QL ҙ3.}-q*MI~sD-~EwsSĔ)My┦"g nᨈ.TػiӦpwݴ(\\{[3rXgٴ5mxHo=[s?;ZcWᐒջZrbkqd{v;D_획;  @r $WHA!˵?T;Z nJOBaTW"r)B<вqr1uBWd#cYRdKtE$9/5y3vQr4FqQP g!#TBP ( *P*:I*- ~/6t7ëW}{ri'e|D Yo,hnz/\vdnCz3Ф OVwNLVR"iG̔1YNbBtL -5V N!#Q-뉩vV?)n@|}}/0b|=~Xȡ&lqvac`w|׫Q=o߼tˏh|Kw/bX0zu0iCqgQǣOӰ'~:%֟YGxF֬ua:ε7,+<jVy=$:ނb_#,>2ͤd~xXȶ {o=.Ձo'Fy El!#~u6.wnV0-KSuXwO,r,IX?O`l3t\k>ݎSJ찈+r+Xr95w9reNZ듁s8?e18ל#u3fqŲH{_P~\`H+QBK_Op~eN~Y{这~?0sN_i8|; [kxen¥wob 3\ȎCׅ= vcn{wFו 22<3^ H{>b#*BCDJN#ä"I"/7ZI@YA9"L-g8*dk⛏j]Ba"Y__&i[Њ#i' Pe2y3`<#6" I1~-y\ܳr]l<Q9Q!jڏv>KfZ-AqEp:LL1RP:937sUgH77]^TW LD4c~x(Dncʹ0:1bvOgZR߹(d9@bhr}8pާB}FAb&`M!]Oh@~@B#+,T(EZCQXU$HA,AHʘ1`d \G#D{Ԁ3F+,&NX)Vezb~z2VaHN8O[ 縯T*XL!ڗD0(CH"XIX Ofvfj:ifJ#$EDX}+EBncZn5#ba$ 杸sFkX"5X8mTpwlŸ ģ#M`*?!qC3)wM&>eydaZq0!6Gr (U4 Ֆ}>ИQnF .$nk ~MY=llfj Tvw0۸`Ww~A/_Ŷ7ZFqoZmY5@sBR ɍ;iEБ? 'Rs\AJvȴ*ahٹ߂:|㵔=?pe0å]&5nc,ø8|冋 1csOCdIcn7-N3ܼ.CL.ޘ$p$iD=&̔NJT† ~#fǀ[2PթF;Wey(fr'7]:`U$p0[>6Oٲ2j{1[-i=&*nHSE3mt:'rp  /Xkc]g Tv̺IFBRh|6"2 SUN?I<۝L@Eo>^ݍZ4@["A_*Lh$IQ U Të)V57iN9jI&5T 8 buOq& I&};R ZZ)ZIlfߤ x{ICCvQ٨øFNf7hfO5[n&}փ2DEQ-;RъԯLz+n'L-gӯnp~$9FR>Y6<7* #pS0t>J3 HfzȡCm4>9!QWIl."0 u%ލqHX!:m$V"N;DqT7QDyB:Uu;fˎO U-ɋ̻т7hC-c)<%6ۚ:4}V&ljf`'3='{1MWa;!m r]uB7!rv_/H[mwC׆u]Kt=KaA<5&+L {- B G`z^CԸwʚ*2-(>8`8鹢>-dP)T`+(M .nςhN3Q_bV4#=pƃqĖ;Wp,@|\Z,9;0G$2Uz46m@:3aR?5[ Hi}WPWn8Az1GB -O_HT_`m f\]Pe1U"AX4C G"2Y"IFyB5O#q(yfj7i45. _|>F3py_,{ Fc02 s04o^e-gvs0[2xp?m?L Q@Ă0a#,5N}a8_S_# fm^蠐 NS5ĩdQuH,9 xCZ5aBS_Nn51O{ߞnBѤ&b,Oȧ+}7M gwUtyYLuFܓosҞ/Qo '£V (`r<3#Wf$S]r)ӯ0F X%ҳʞe2 0? =ZQl ,L$CG_{z G早!zu%_Vn4\p 浢۱',./S, ~[V/KxQ2+XJe؞Mz`8+UŖ'`yA0"jodz6Y¢C|Um=)Uq(;)  V%vDjC_ qX"S X}㒿INزa4V0-2-O, 1m+Pm3kKcLV/A[ϋn}ÊB0 堞 z9f4M:SLJ'D(eNCӄTkhP>;O۩ш`2`BJPFHEGH$ "KPFH= seXGX~{wo*,[,M O8xB:wH~A f|a4><\_ /?Kxi=ZbSv[qI..|K\jҰ593Wg)UyѸ$?__8|uk-ZAkԾjsɈ4S&hp,"/XFI4{$$@o.@Y\SȋeB eB=^TZa`5)5 T5?=(X_sz(X3>BD. ! ZB{Ԙ dwkXDdŶ q{ܣ=H\|UMV7ܥ`zGqR3g~*.WGklC(q}hfWXD]L {#ڄOذ͌t"$ˋ_0 JW.EnN֓y0xխmYy.Coȗ"{%LќtL·ťCa8Sj\ySC)~l`ب'k[_ DIoQ<ɐ?a>Q¢Ciy,EkW`Kzp~=} Oq k[k,,ax1+IZ,bd /j %'mlݪ:/E^*HF5~jG4nY' }Oz|u~2*޴~ʓd];H[76N[o=P-pH"ڤW(IfG@h:`< W\f<.}pd^ʎxe_cg 2O z5{7fܿfX-pV;9J"n >~8 ' `:?Y8~EQyhwN'%ޠGWmQöj,_L:\QvhπFL+VA_n;/A4˴ f?&8ϤijѤ236cPdY\}<&#))Wq@c, T뀲431 \jgTף'0? Iѳ<6 ,U0].!w(K9jᣦߎ% v>]ާaO 1G]ό+Le%-V>3hPņ\ < btK  \xmrpE$ dq$ 0AȒHrʤJER %i &QBcB)1HʔgD, >奭r"twF.uȹځYqVv=*cʜ8nۙTsn!`dO4/ ?)_>W>,pTe;{OŕF1KIDH .h#%X"c,8gI&QȜ7?-QTFS4 JO'fiÌH9.~ܯȹ(QRg"V: >fOQH3Hbh:B N$I(FYE$ GZjf2@<ԉD&qJpPH1f&,rqZy3m4%em t,jȝНU nt,L}[DD hVY$Ğ4& %M<MSP9`)H[&4#ҔJPnS4M`ZnQ3k7saWEIuqc8\C_p6_*an;4D*PX?kAshrscB^šv~9KWmݭt_TAǦ11r~Hvy9\c :>ț瞱h8n+ r;~[nusݎՇ*o8ZsCP?N{O1HL"|fCQvo}(%椴iXkDZC#J#*G[;[7>x[o?4f,t(> $ڳ)"9֘}!ܱ^Eu8R;{v4)2j#&=d*Ԛ3zL3[Teؗ$6ѓW9t%+y׈ppEj4Ħ6΀ҙ"6Ӻ+-9V7;Tc2U,E*" iIducb 8L QeF cep LuΪОjL%ݘfkwuDIw ̰챙,꿝?]HfƉk&Fi!MDb&#sFâ,T+NJ5K}6m-.[I"PSWJ\>-{Ve5ՙH,7ߴ"ϯYl~΍%Wjto3/b=>E,׻=?_c} ((7 6O"v~O.[eM +i8H9 ~ dD aFqLx MWu(SD{ſدF)4RXRy 2 R3:LxD@5yb%m4L >KZAW9yIIPt6F3Z tӘ-W!kчOihv'hPrVǽd+}hUV6 FPT}s]_ޒXM4v!Xi !Nw)IPCP!nj} HLDޖbM2I{t~ s0% &E5Yp n$gU '@!{n؜ a F6Y}9D %v%q.BSv$xme"\(w+0d}zABYWĜ!}"RSJ[o:ݰTJ(kyUpM zX҃_G#tе6B]r:#pDJXL&dF$FbEB))AfBѿ4sxGW-gkۢ>|a>$e C6eu&ߙp FO-mr碂cC~&-Joq ,L$E2E1'J2QT#8"NPDC {VzWQX+iVbܐA%2B a d$VF "( "QJb2h3hL0t%Kvmx&*R*eQ# Y$tCvWJX$ )P(C*}QŇXOIŮ;!(yuޝs:H^OEnnu}z u{AOG:\=Js7ۮnO|x DWJwbe@A1_g˱ʔwYNf3^շ0pWYZd;xWVMCD`]S4QQDžj'T3BpPR)#mtth=zS)׫VDzʸi;6KѡԀж 瀆/(8d{y7R1ݵ49ODv1kW{='_+#,eGgވP ޼]]$Qvӗӻ!ݟzTb]jmLfeQ&nI=T5dQ删^hc0#I #rwvM+ɟ0t"F&/3sshgIl/E5.j85hYі21 Wz:(W*XHၽ?OF}vѓч㿍;|7M~ gJ0ohm BU|>Ns{ `Z(XW (؉ϑnT%wjQ=HE>I%cU5"f4H8ǯ\@$.`iˈuwOvτن'FŸwvRA\T@PfwFI{dAJX,b?>\>sp]..7q~U$0RrT.zv1@W|WRWK3D)STwBe;Db!=joԲDN&ɞ]ܤƧӜ*nΛè|j# a–O(q 竽Q3H୸d+p}ҝ֘2К~4V|̐7>Ek)J{]/ ,`J7iI 3,bh #Tʈ0J%*8#RTmei]1J@(MB,Tn(F1P IHI&:o0uHC73 zԆ> fX7]4z`eo谢Ǝ;<^`aÊЇ+@nY M]QR=T_HNO01%Q%2K$6C69>}p*#Kt=H9wL 29R6K͉3_DAo6EJ_e}(H9/J2>+} q8k' u$'~(0Hzn+sAÐ9D% 5\I?[ks7EKܪoQ8sa3wv/=%HIE$2(A&3BU$|JK&Ϫth{[/kZU!YdA@bg:^0®RDBrޛS}Hi!kTU]X{i~uU`:9KC~a1/ߵZ6uE} gityҌ(6v^A"= oyIx; /?tÇ%窋p2 -|oq<[aOV {&alBD k+DFgzGi  ? dza.Hѳkwǿk22vE…P$Iq )dH2CHĘ:}q=Zgy 9Mi0e BȲ$΀&a#00 #uIԯ(C-NwtPZmdAȧZHu mHb!n (T  +SXp աzۑ0h>D$iJ-D>Rd#D{Es^AA]LŹ[JnKG1ӆ hL&zNd^[x~v\TC jC6 $$W̵ZBrRAbl$"֧u;y*s0kbZzTѫt}.n􀮽*GTx_;jޗ̑t402M(Scf$T0#Ho.V_蛅JNI{F- V"z%"I".8gPn8EsX"'V  n?[%m+3 2f'QtYA,vȀ]<]a:FPϥf;K&w8Pm!Tg`%)KD 1GA% ")YcqϝrƼ+ˆ'rP{~3`=dx^`T ,vPqI4 $ݑp P ow4aka{ "}3qQ 7}R.*ҽCaȄ"Ya@0X& & i̲(fTZ G1!<)N8LPxec 1(T'! 2DQb$$NHQ$d&,!R2 Nud37a(!Y=0WRnO" ءv{Os_|6JvvB8Dql@{B@N0C'"'LZj,CWH>=_KNR0݅Z8'󔜖  bBabTL)6zhťNbb<1ThG Ġh T 5';iJ3YNMu'&``EObRɕVk58""`A$&k$dиb΃1VVy+9edC 4r,G ,C(:W hރy 05ܝI,TO Q*q_RskI BSW cłkĵT3%Q9fHuP̡(eU ~40+  ,JIdiOBB^5' i_w qHW`m,g >V܊tu+6cmV[p7Ri/8h wY*(#˻ʒg==x# *zb6hu9~1'~MdRY+`*vra.5NˣBñG&'Dd> Y?0ƔYNmMoebC_-3hMQrζ/2Di^'^ 9b4\XbeUۧߊ?*N&rvw4>lx˧nfS={L]h  C7Cצ'6lPͅ4] !n.WoG;Wʯ zڳ *V[}1ظ_=;)#ʬ>"$'Cb uH,1$-R! #;ܔ%WppZg̚#y8hfgh{eCy↕Qcf{#nY#Cb|nМ+t1V#**ImDEyL3!Zr(Y C8 !l Vq.\`R&)u:,_-  ]ZG)J͍A HEKN1Ja,JD ,jZD4㔾Rޟ @K~=n,=W EČZB[Ab`*`&xY$* e"q9,xe2Z5W|`B L8EBPr>?bO ?#1 )0HKQY w v>L_)h؅*Lў<{Fnû]RnUC D]`N(T5BQ&D8Ju.r#*TpJލRikQֽ;n;9{1(eejЪV2{KбxȭVF ;pPUm!lK6WI•)d_pB YզD7tq~|=Ƕ\rq6j>8K0ӤrhyeҔ>^w{h@}M+w~z+%13OL/|zx`ZϰnzEI;m%6u`r[_.a9mM}Q"<6|x+=[QJ$kP׍qG)$&Jy#:c4n#&շv+&4Wu!!O\Dd֙VU~ 1SœLS 7)M]2!Xˢ"mfal&QтG&RM Ex,"%&b͚"0̄D'(eH:mca"m"ʕ̚qÙҏX3_]ꔿɒ៪*ؓ y(Р%7@BUw+vc؅<Ȋ^ h`*ZRB$䉋hLTÏF4[)rDt&m2U}׊hnńj.$䉋hLЂY!IVtF=eW`z1_Uϙݺrqtj Sy$īy2=F;xxHacZjQj XZ$"@a@;bغtg J$F` kM9h5>yefgT)<ȱc4q5G4惹L2].$䉋hL1*&oIv5S Kt hS=&H1ڭ y"Z"S{; =4a)rDt&m(B 2&H1ڭ y"Z$SnG&ZJZ7 G878EJĀ ,uFcͭ &x2b^Z#ZfJH#ػs( "N© +4` \p Nnᓗ7;ֵ˾GǔGZiєы†of@jut6^Eco%չYO:K~Q!ht3mځ 6r孯f!)yGVW|M7LADTT7C_;\V̏^oqwͿWz~gqſ.4]#4˛o^* 0kw~aw2!SH$unϷ}uE>vO*n]wq̜3X:AM$n|[RNR-?W%W|ߟʕ׾}۟0lӆx$v!atH§׽ !b Hړ1 iHx*=*Q:aw d`@諤S[R;5*Җa-=|"\HvApj%Dc[?dU2iW&cY:E1mKGnWKΞ[iZ]H(&CJdP])[jwR(cNWJ2cVJM=LtDRT_]k9uT(g{PeCґ .f􀘋tE15ӈǠ ICFdMA?|_cv˶ShOu 4p?ꛏo|muwKYO*U&q &vq69%Ҧ@maoU}t|w拴IŖ+~Ҟ]; n~SU ~;|qSp_r?Rz.N/Sv#~CGwGx{>eH%*MwyKmb8gϕ/MP( (=K*c<$.g.N,? @^~ڋTi0bX^|zp\lj/ëfy|:WC'su㫱K=O:&>/7tAuM|Vw/o('^$}缡 1 >sT+?OZs\O z9HD ֺw(Y?V0`/ blnE]bC6+$oHM_o/fKqj~78<ۋ< T6؇I}e&ONg_@УjQgӼg43WrD Iy^1be7"N`~8"k0DExlY3 _7, g%WtYQ% h4`.@3eS^>nW3Y*YJDX_'^pcn/Sx;+{gЦ|dۦ=mh9EAm fCHyg 9@^"|Ȣpy献 8>(փ3uwIy'~|[h%@uC0&v@.aVdڡDzr_ N&KXpn0>_x O6!B U{xw=L} jiBQx|?VOռ' d ɮԆڹfONe {U$6ffybOwf~sKx'Pt;DчSts8KpˡQJ`9%2i w9R sgYQcGr_QXS8kaZ.9q:;B&6jL8W+#&WĢ\X_@ǰ%ځhaxC \-DIt#l ֘]ЄRJ!A8@gHe(WV`sIxn6DF1 ;nQvب16°t=% w+&("P qca6XiPl$HȑRX묟q'O7jԼ۟n+/vK-,>ZzRoP޸ܺy3ؐ#MuHI*t-6Z2,¹󝹭ꇏ3goJn䱏)&U BMczKUSԚq+&tE{)Ίd:1 66:ϼcrU K)5%eʷz֤%{Ռ8k+aJ(]LJǒ?Oʅl~GDjWz| K0uKncI33&kve$c #]kޞ}yy9)ې/J˱"ٸ`]Pz1Z ϡ'ϋ5ք0uEӫmbڮr5' %~Kdy9Qvìm"i.bA͑R{1}'M$݃i,fWهOiQgA[M.*u?'na..:DMŰVO{Ep3i%eaEY#tΦLpjb1|~?&iρX q֍FTQA¯ N/yȔwl Ƣ'Jwp8ӊtWr 5kpR!BT-rʶ6x.#q͛}6u5Ur2 \r᫋AE /sAɖvS>$VmB9X^|AtMu\i^xqۻHL`?C޼b~F(b+[ddF 3/h6_Eݲhe_e/!_CR Qt1oď `5C\͊t dCmU'%"w.!(AJ^\[ ʝ/kXp>J DDGKrHZu>$y6) CF\0z"1uR9RIb!sl@Fk8r)U|#b'׎, E3Ӻ3'bȇ`bR랏tkEj$ pHYsIp>sיt#lT ?لlsT~`y8k# ?uGO@ ,i_e:z UG#(؅׸@sd/x.bVpvbD"6Z&_@MكpB1MM S]F7L? xF%|F&s؞ 0^SG@BA>@R+50TAuS&@Ǭ]V,^2(z]A9DV. 8=/Xڍq!ac3Qt#B+;i0'۪1ImU=;a\39?Yv64+Xhdƃ܌Ye[qx/\NO=]]n~mj?_\?fU߰m:u|ޏ'|Wǀ e<2<M-aE_(M 1 N~F7l>3t*"c|ߊd ^[pTʸLd߶c'.$LR,X(Xr1 H}btTi0E :X<BX^VrMб0IA* 12bFb"~s3z<G8UoB?8Uo ;֖,ׯl}Q^AՉC!֥WQx],0Kc^h`T.! kVny$_s~j017,OJyOO!";x\.W߄狏KLUphh/`? nqr< B!4:d8В(ڇEF?/ 1\WqrsXb/k5Kᒬ<]!/B=ՠXgx^r# ͡`Y=8,괢 9E z<..l9n`. G&9gܙݳwA.?f yq:_ĖYaa\&Dy ث, YiyIw;ɲ{Lk%zg ܮ#5[돏-_{8grlL85LZz{3I|4\ 60Iu4hrfI nE6 ^!Ws=`57#/&T*5Ž~w2V/ձEmt%cpkLjw@ӱyl"/?(\tMV ީ)vLt>ئyX@vodM£|6Ey^V?;yqcL;Wi{'iYyFTDlchomjxn֘1l`!xύO#:E4^N18=:CpEcz v f58EH3pBOe1M9 8\`[-X6 \(V MoeB*&n,aU6&c7Omm: f*mt;[uW߿O]/\%.\ &v_Ny1"'_Eoc(標]+Wrғo F6,'w'af  05ؖsgJb>_ë2WR\ 2m)tf!*Y.׍XP҅x6g} Y4VťWPyK+u ÝHWKԱG4xy3?ٳV)N86G%ַ[j^E%C옵ĪMci0QQ5P;#FׄX_*QR #%H,"S2P484aBHZB绔ɉfñΦ}6'Aq.,o'`0# {MrSYi+¿7ި_z~Y6z>pDB&ba4INL0F8$1'sXT?|>LA_, y2*7Tא\_vyo *^^]X6Ӆnk,? =6nyUAe9,POEX2!i(,"TY#1q*3LjR&:K"QHv~H+=/u68xHoT9AņHrVsYua6и!((M,^244RR_(S/FQjKxFSXB ((8Q2fLNNG" IDH AUK7r?R23_ڢ3ԟ[Z=2Yl~[_0>=yׯh>B-.ŽXo~wrn{E 圖?cOo_],Wk>VS7,q'YHv_rJ_MnK Z)o7&/8Ռ3l 9U7`hW6'$,ph ) zw6!9R$9aMR)G J#M"]p\2` jD|Ձ BJ83K j-f$ EV|aic%hǒp-U vVu5ٓ)i_;w9=IbHc[&c[qM`'#[k_BjM@N[HI. LB!K (^a\1c7@P!;U]LcZZmy>t }R FJ2j :!%=edW%D~=3(  5*e3V)30b}^O:Mw/_?H_.yUͶWP^>>kSsC2`qͦZ#EߝkS}to hMq( u&"3@ڝL (T(fB$,W6݃3&S Z184\g T\ $kSq?f2$)8%4Am̘8o HibJ_uŵYeS9hf(>9:`#3F M*UeqL(Iyk sD!4Kr@M5|=T 4E%$Mߠ'fz3Fr$$ ) 5DCغT)6Ḿ!Mi6ޤXlLNə[bue M IH{b)C! g,GRX*āDKnPmզsV}'eE"5dh[ u&3AI! &ZpZh>#7x42,c d `;TK"|HҰ"S ن%BĔXJm{s Yli;EPZ$?&d$mp0XlP*rs,3}z@^'ɦ) vrORUZ&Rp^8x^dH=:C߮$ TIIw8CfiƐc7$THgeOH\Y LZ-[ E!g gЍ [RaE{DYխJE]#UB?;Xm`~/>E EH@qf r.d.bQԋ,*_|鯯rI^/~dyE,!0E3A-1MtfVW:4ޔGYAU\[b^d$(412vO_;^6',Ϳ&{hݤ>[s bqu]zs1l{cxszV|WN×!Dj(A #X+u+0r糉8z|4~k ,q򽱊W7\7G[]U}9U8msIK| ˭mz]5~|`m(eoιИبW3}?i9TKgUԟ{Aֱm9y?RZA= UTgXkş]Xۚź;_SLϞwWsJ}m [:GvM*'JH?2>Āav8Py<X2 s; Nhh" ~(}'P=h~,>_B\ɏ&Ҋ{ c312ĊH+qdT/mKgםY=Ct>.B "uf} -[+d18Z=Pp(:x!**Sc"CH{v'¥Q8z ~T0Zz/#jƧ1K@pk]9NG p;VQP=MphRz4cX *wۿb;A۔04H $0"iĸ tD˒?}lI Z(ČskՖSIo6gGSmDHac[*Pˁ|M!cD%ѼxX0;C2]+32} } G/k+xTl,TqL1i W9ֹKѱ̱U#̧D>c_Cla5Є҂$5$Cbc# 5@i-98 emDa͉TRG"o2t5$I3{10Y͋tLfd>s]_S&j.>&Y_fr'¾J&'yRahS$CR^&~tHJHp+?4 cyBo~9-_ӅZ/}urm+@F 1{B2#eR"DGgD*L=>w'/cT,m-L~_ cOu ~u~ۏ߾ |5~'tiL \Dcc4P:lx$1Z3#Cj(EEI"L "5o༆/lf4ӛ0zJ}>hFGhxy՛\Ϧ?O3CŊjC*N;,|m",\hOvD:ru5F5O9z|֟/YND+{/%ϡ#T&fo) "кFɼ$DS1ĖxFa kd=eyJGSgZ9,)D /*B8> D7pdj {) ʛ%WZ)`k e( *`8q"VI "5 ,&@jpc󜝨#;qroxp5ra&V d{dMY=?++Qe%6/K-9il^I%.A֒J@fν6\#1 LOe  8)i(ra8Q*Q5XX@r%j ;DY!`dxYo ^ X?Uܨ[ 󯃇~gJ͞IC)6Ð'NNQ_n\u+ :}b%6nڶu+huC8EK8%gR[)9)c붿 T#3V<\ Z !cRˆ[5'AXɤ. L9p߂786bKe!{`@h`UP$m L!Җ6bO,0"75r!DVxC$*68' A,REXB%z?-8ZsS8 tt+^z Hu ,T/Ɂ0}UBP+`AA6@X䐓`? .P)`r{,YeB: =s@*rf 9458 ̀IB/pH{259 ?h'Dxxܞ7šN+ Nc* TzN|`^Wór6!9FZkR-(j閿s[g79Z4Xy\n|`\7Koօjb.mE�lC4=OfUqUЋ_;=3\`دW>|> $LR [w4StL6a\nP,Qb?iqw,Grͳ"f$Vp$ޔF,i@@hmlCpnǷ>]ZG6gla4PjNTXk |Cw8ڋYwzތ]fU!:rdI u8_A 5 7iY77ګ^~pb }SM䧇kU("=tw޼Gu㗗7'L*~1R;b*m|=øCYuBs\ACP=Gzw`z!]K廣<]_7eJ;jê6y2D){]uޫն.I|6Ż$gۺLY~3[o|nؙ[4vJh}T㍙\K (зlsƊLptp0~23o!-Å`j ӗ6rk痎dp>+f(o uc Y_hQ\ +N~Ot%0[#WgoΧF7g]NL/ol^IGHS9+f}<.s(wPuv.4M%zX4w`Ev+7N8 9hZvߡFɛ}`x+俒Xnoăї&;𿅿_5nΦ`RW2N̤nM`0QyvĽ}mVK\lDcFʼDQ' d`h6x:j+΂s>]\]~80_!l Wt.wڇ>3H;w~t mijly~YQyOD(}߁uh2grU"  !R^y kC!Ct)'.NDX(ĚFɠ 8+XJL@{"c"c,Ÿo U wL(ZW1έuU̇j]-҂ĹZkuOhcE$L.4GqDK[!%V@%#*dw 2xf)ơuR2LF]7r;!y?l* ,Վ{Q0 Eyaûf{O- ,VmkH)؋C2]%?%ٮ{Q0 ڼm $_zzB'GmЅ(DkG"!1i< _[v&"a٦N* 򇆳h v|w`|(q1OT(FÕrŊif+63nUVjiiq*iޗK12LzCиgq$ HBD t*-"0|oip),.mk Bwq5Nlx7~s\p'6Q{4 \WI^1ѴWHP#"]ǢC.K[0ڝ⼽4k#7(K1UD\zBƽwa5qWϨ,>;Fmo>X,R˄l/ wYy"%v̉viG!`U%TKӫ /74UI zK9ݒV}DZ Qd{"2e[5%U-9[hϙ e@KosAv]?爲!vƻ#cew6}M4~\yF7*NZofU)5&0Zޝ>/@(cСsCHԿ~J)YjNl&Jf]N>BJ5eٻ6$W:`v,XZa/@T! (Y߬@4Q>J8F"Ь2++3+Y>l; U5m#MUB,e=8Bߎuy,uKR-Yˢ:{eA7]$#IvVF\2'i5Xuv{ "@5S?9: ^|Ɗ`L(R ;[zJs1cW0I׹{%fw% J_fD 8ymIg74 mlG}:Rho|FG}qL kn6XzR޷kKΫr[9윒Xbյ|:ڡЪ!$C*v5!D3dNQ0ZQ/ɷ0+{^QIyS2e|cn첖[4U/x$Ϙ]QÞDx5w\FIDӈi4KsY4} Qy9&J<'u- ΒNAP}%#dhD ?i] rRD rҧ蠺?Y!>:cL[ S{RpmO6t NZDq2Bs_T~Yu;dD"5Q 摅קO#$?QjT Iٶp"DmBغ2"1>mZ&#ScLh1#-}aJ:JS]oJҔ(g,YdBf(k+ ҀYXɭu$KuS%ӂEŴR0V MP`6 5 в0 1iASĈڐ[MK]q=rWVI@M'fH)8ݫ8-fZP`l~13J vtE}7bIb8Y: H `Tf%Uc2D`cevҎ1Bl:V,A;vf -8Bq- eU906T'NGpm[B[9$LYOiT*ug6.tע{r3S ctmϮa;t_ ?~w/V꧲ qglƓX (7i]gŚnS>f,{70#Jjuvս껱ݡ1Q) O%(sIc.@d38I5B^H6S#ë (p@-M] VZ0IjH #y8ik?uy<.Zpx 95ov9{W~{V/B{䉰(z&׎is=7V m/Wx&ϤpmZVL—7MG-KUn(C(:0IxS ]Z<e-h+Ϝ29(GFSASo8b=hK*1:(N)Vv@ (3t eDSәt]U=pQ6mkԒu\xtJ>}%FɯѨbqu~~7 ] dwfz $27>x{ry18hgY۝mF8"ZUghoh>ɷq;*Ͷi%fKj.P)iR}-+#t_3h7ϐSl*:&$I W5!9Jp }v0~1BX"˵D ɽW> L_~v;-wf~:;o甪8X'z7H=(fsseoJ v1[6ꇉz'~ܒ;;ddܞ³P3׵$xwmYG5:2ԫ+=;J~qZ>VGYR= Y_e_R[/bԏ2~:LDl)p>v}tTf[>\vQ>vu11{e#) =GttW>o ln N7WAۃ;Nq) Cq邢Sh 3]u `ayNA5Q]HgbN>[Z3(Oh3Ti!VQ wݕ7 pN.Gp`&Ӻ٫I3M]piWhۤu0Ǯ|,;IÌбVw$Ι{O(9ťh[M฻ƘoDH:S.p:#rK&Dt1SVHػq}vf(OR[Jgo8(t6׫Rs_˾{[?50L `ҷe*P*"B_7#jV8|rQMm1Y0f$ߧg0̐)sPK9tK.k*$ 7smUݧO3g;^>(Z$epQ51%"HО|[U0bQ?Qpaqv@IONI@v;N x7︘4N~=m}aKΙWc}7lްd9u+)QVtK_M`__ %G#G#G#GN'C" ;B\0~V΢qXclegռ-?Ggc^^γE+Ms= 27٥ ⛸/?"RT#Al]*x&+a<+I@LS&p4,'x-h2KZ(Tر7P!5b,^m|FrNs]Y hQnBXXJ-&lc0%d>L>\r1#% ذA1$S+O0Cy9fQ$ZPT|ʩ$N%BiR[&v=":Ą| D `XĘ# r0P⁓sb@b OlZp" MfHK@ZLZZ]A^ֺJ!a2vC煈֯FrHPu61BNYwEz)~rh O3iE¸kJzxW^SBDž&$3 _y#9pt$F[yHql#w'/{W7Hgv|֨|/Cӥ+zh[MpT"8j]9%m{1XxM N "Ni3.̇ , l߲%8:}&{pttpma(kd_An[#Ti)kccNK-kJɕ+8a`OKkc\vטJ>@xkLH>?>zD#zfstW6xZ%OM!L>rHnh-N7e4^hKOvgGPu9 k׾^1a9`kh9BA܋x7Va :/BSw J[؏&Pn2:boQ [6#>kC|n):Aݿ'1gI@W>;қnY^Zt~4p`ΰTAX=Z8-_;nrRvˇ,X%C5<8LM 6$DŽ'超T+IӨ-U5,:2=-1qlxB yECSDڕb(^%ߟAנn>?QR V?v[X>)g-+,td!߸ؔBw#n219x]^ bz.,7D}Ĺ0TLл tbhNxV0nьz.,7цM05UU:+s oNn8y?~,3F/NU@ˇPfW3I/WS_|BP&r5wyȼ <ff~nQ(3#hgIf'~~\ >-RMԾN\MtϦwӫdd+꧋%&r|Ve38oٍ,u[s?-/7yȓ!r%dqDvw%lL1Aub4! lJd8'*0Fo]p$ynt2L\6}= xqyrU ytE6Ӻ';)_7 !Spٻt/?X5s7TX4M;/f-/dz‰Fw[iM'f3zVUKK_<:-ǟ(Z au'bYn<[ 6z5 _$XA]-+(t`[z@mh%ƺ\#|<͕ $FI%#ȇMf;r 3%:X $UMRJjdn޿%DIDMSH:pyv}]Z;TEClsXj%u؀5s`ցWZ[+hf~: pʄG T(3nC@sjo * w^|N7::pJ3k´ߓ7}LY^N9uy_]Iw2B.g~aFj(T*xKDA;OLǒ4AcF{<<ԱhlP-jÝb*J%;R{A$,CwXf)` {ei|6U4e߭?*a* T?.VMw7KOƣ- "Iqs)nrw{EĔI"JsTu?_&%d{zVUJuKdbsf\CLcgnJcܱhc:)$L%%}CfЦЕT# *pTL8R])g{xW:{@^K/i2ԣFf[7@K`|F1| %w$3aZnvtYue8 .4IXÄh]TlK٭u[^]C];2-w$]Geag~6#"1v@o(AvebhdoOm}RzР䇿|w;LEiD:g2TZ)X$ǻw[>i1Mbv;K|}wqAϗ n[!`ʼޕ cIasT ?5Ѹ&u#5#9;u[6- C{2Gz+?0 :BHe7;37VvacRkJ_9:bP4Q(kd$}ɈǍe $,n؄-a}i"`&l< MpPqcl~:Qd DC !:Y Odn \ ˘ޱ^(TިVqmW;reBfqB %{UHr5ZbVvZ[2$ke5W)34ժAcT`;%4ݰMjUhUN$MHV'U2:eAetꤢbXFdqwkV''luB@2:n'e]>7a܏VK(^ ,yk@[b*X[|>O#sA_P`c[eJPzCa~'I{ y}#ʇ\').Hf#OyT?aO (CdfvlLV8a9, ™RmjHq5f:!Nsy!"_* ̆ٓ^Q1 #ӌLg6g2nQ"ɑ 牐_!ΤC!k#3 A5*D%#޳;ɑxRŞ8S̒^*(u*GJ"`=(h> 1t.<8ek~m55 yvohvKFm5Vn9(AF.6 pʊ ޶(}t]fïWw'ܭ)8n6>vӚZ YQRֱSbe.. ~_SrO%eۭl%\==m,[jnw@-w붫T A%鑷_oxܒF,/oN?.burlNV69Bv;Y^_^N>&v7|io9 iB]6âC7qI,C>F)ɯon'X.,KS *pqd#Æ; OV%JZ(cJ&%`Fmo4ډFLڡ:fX$'_"*S % (QWUlGO =-d̓]}_ԫwfhm4e1C? Ha_<ĤojT&ˇEܲj# !y~Jd`XDb0l zb$LȁvM"lW1&5"Ho\Ƌ2%+zBL*[#-|ՀrA )+vͦB>vDV-j&=(jo )2FX[NN5#Zɺ-٣59ך5]4_sDGcyI*(8=r:OHl9O5/y0c}n4, 6;ƢڍLKѶQFS|0Ge/'pϣGO =m&~DF6ۘXPM\̬uQҊibF&Y2IIe_k;3d=. . wy֘v3S>i,UN. _+JR71򛓓~֏ DģUjTwn!Is5*Ɲc椽*r99NeJE2^BtK%1w$O{1+ 0YYd`ʄV˽.# .~>=A`dxKF)|{ yEѡbO|xDޚ6^*,VXhX_N395!#kUyeok-xq .9_r$d  O^lv@"a廻txW)Ȭ?}?$Z΋zߧGXǶmL>fl-V,)JBGKcQ˳"E34S}td4tSj$]߈V! {C]_vk׭m-@O9˳FX"KM2J;`gCg@JZ)#XoT>}Ǥn헎ov$hŌSvǤ}x;{Ў3FzrE] b:)BQ=LհJVWF +xD2ϊAX"02~)'ruCAz%Ӭ*>g#Z<ϖd9P1#1ÍKJ5$ƐE RjV^iR6DWv8A}S͂Ɛ ^yd\39FXr#6~TwBgwr{c 2)(4p0Vky *szL2H"I0&-=Q|A` 3f˵jPZoijE`}Z nSEvξQ>|%"T:N!o'm-[zYW{Uo[K{>WבHswy\]xztn^J0QZa"j:#Y޻Bv@ݥ3z>6|kG5CN|FܺzͲv-cy|n7+KU:X -N-)zcn͋]n'*o7WOu ćU =7t-rӯmb gQH[@'| ׇG~c B~kho?޻`G[lSx}TY) 5x[`P1^7߯jUQ%rUJbF*HzۈS荮_%K˶WCQL ejCE[*׌Y_*% N׫MkjbbJ#\?6,]n{3Z#dž3S:qBڡ.`D{;*+ )$J EMbNo7@s{WIgժF]loѠ=OaVmN]m/eEbK{g-;;ߢA{Wo ?n>%Ǵנb v<9U mwmmvwE{)_ؑ>/7rNr=Z\oq_QKQl~6o~}V1۱7-wGuUJɳE>-R5QFv4'R{\RTum[>@ WD' %~Ǧd"Ec3Ԫ?H>IHʀʠWV?:IUXiii#Xoi_e+2A.J1F#$r0|p[@['z,-[/@^JfG\~^}NGiPxqn׉y0O<@U|w6aZ[}z~qI f@ڸ{#b\)5ԙE (SUFR@^c.=/i3im [ \E 3).+QȬyat;Ǝ~zqI4p\JKFy~a=͹+Ȥ#. $)%E^8Ҫb$/;UcXNlrbMSq1$u|R=v^eTX^:(L`Hfuwk5©y]p3|(+}R-M09M :m!t..:X<"7$.#PW8^Ғ /3+z菺b FER 6"Uًy65?b;3מ,4 ĜhU-bU^mٜfzZ3H討n{s ,Gj{<.`hO`FI&8a̔ ӓLP'ގq/{5LI 򵤌ň]ps:G9Ev~UZS6B&'{` |-IכlSm^ߖ08$sSWW;\wP)YT"Evil!eLlY(=.,S6j<@N_fiV ƂSʁ!nSBˆd/t4lh!,V+EHNekF$o"9-qX-ۈhYPCW4ƀe􀥔"TÉNgcL,<76v U^B;l[f+HM#|jZ 3o"EmHM{Rz /튋%o՘C][U@ܫ()h!vV!jJ ZZ T919,2]̺܀ opA3_*U 0vRոq;ξVw3>yG_V!4ky݁3/]i!|_տ~ٌ'mf SǴvԛɶ:j1jq9]#-W =p"WTZ.;Q!oN$?|4S;'wt=8Zs{ {luq[4(ތEWV‘IU-2rF8Hk+J|Cq.jd⤺FS)gTʋJp0 AAN$e&LUJoҾa_"yDx>:^!JSh%BfKb wDfל>#Uc܇Z5~یtNhLkKYrcJ7rJ׳R3~3~O* ="QFƉ ,Y5UfM?jrcַ6ayl 46A<ӜxnwnxSF,5nt/i 5r,.} I#oVb82} R,6odWU|A٢VQ*[ Sg2DKDҵ[-[ܤF4#94:4Uε(:zl*0>8"86&;Jn;`]9*+PfXS[W9;+1vZ r-f@L)Sl]+9q  IG(6RJN(.zaLUr cCyn3&FbsDyֆwx҇2R(83]H1WFd62w3k7*XlFE1 0YfWooD(!EgYlj:d:YKjQ!LIQsc@M;Lg 3 ceMԣxY=Mţ D> Ke>>fxx]LwϝSS<0_ 0_ p-;.)F*yl0?T^+$)طF}[_"&QOFQ<)#&Vmjt!˛k`(ʑuP58U6; 6\F ~ķRuEXiKR:/b34*XJLM^c'Q?0' H 4x ?˫898 OiN8OО;8$=K.P9D㺠P=V<+(Ckk?znME܎=Rw;A7%w7F)mE*Axyի  \BaV]\_HC'Inj^נ.Ul.C)M[.7uPAJ%g*C5e.Hj@5{b^\Di v1>fei]2b? / tF-o$/\+StCg Bgwqiܡ g"dSIK͌tӜvhn3S6G矗y0u 2/ aGۂМKUya͝gpN\eR+ƾPGn[k1Wh.Q{4N'٬Ylll23v|, Es+y?2[\TQ /jd1W:_RTˈ/:a_i܂d,v}I ";(36(jXxqԓ"YUv: dԙby2V 3Z/'b\ʫ˄HYß M&]]^:* Ru^p *J0bDθ5TRJُeq eL~ % !3F"F</wZ-B0QȲU6:lAtO>PMטD&0>ЃtΡ9n%"GF=ёhNGۣ5J$,0),:C]~{-ۍ@%‰_c KOAj3mc01|Kgi6t+3${$]`:y B!^G ySR[ϩ>\pTÏ=j>vTTŲ 8QA Zo8%,aav/$Wm!4Kel|& qxilvϤZW䷏@'ߍ[Φ[Y |V 6 Uq yNiH'ڕ(+bJ8΅ȜEn/*Xa3UEB S:n-Vy8ƣMQNB0y,'ڟ_Pzj3yUf:L{o|i :vQMy. eRKbzP6)-gEcUHӀF=h"g8s^k%E΀nVqjC-@q1l}{P-fc̕n 왃(!a5b;;xX-crlO1S4w4?r&Ǎ\kdɴəԊIߡSRSmhUhbs%sIB ( ţK%PBO)n>?l6q0}姓'hGH=p'~{q?U O~Y`# ;0?FR?m7tB~,aٯѠpz~qح?㗇8FW#-Ϫ in]v %~eN/KCiLtd WJJ-.gI{8+>oAX#Ķ!6>-GR}5pHK2꺫J wMF舢pF"Up)ͦTM|a, ם]ԺL G [J#Oqr- }[ Y>Z">^c\Χ-E A;ٙg_}*~8r_sPgbhc"e ,|/3ZŢc{ OeF;쬅*'֧ ,6>.KՒu_ %~~7xJL>Wy0Pn%[o>|X+hRYI>qfP<ɏz`B*(eBNt^-'y]6GNԐVR(Бiǵrr"mZz w`)ڌu<$=%X9GZe(dh(k(ْ!.Q&(.:η' ,%(plQBH:rN1^$G1h KtZ! )I(d]w\IT(qLrmxs09VX! Ȝuu3Ȃ%WƔM궟eN;"ЈrBgGJjH rm#>Sv@Q Q)u\8Д05/DQJg)p2bӂF[ERZ ;X.lW3-B:fZזOW[)OwG:#i]s)P0FprEГAw^m//6Fry$FJՏ]]j0% tv$Xe)`"TSvHl~YA!oL<D4 -878dL'´!*f~YffJ5<@$۶r`6X4* 2$Kr Vt-R& l`Q"r! #Q Y$ )e*(D.V&l%[hWy"1W" 5fD)D*#H >g.2P'U2X c!"Fڃ X'%E@oHbK# V1CJt2`$j 6/]U0fLUʂYd)Ӂ x" 0 L ;LbP*<38Xp'8™*|8.bTS1vL)RST I8,;8zƮA>F(^wzo5/k!j6 tEL!-;2=%fͫ<E*74vNynD* MAr$b x !cĔR "* t}/&54le`R^H7 4(B!\~4񲩐pǥ a"9q }l>rmw^I'.5Z38 0?9޴Zwv50ФhAt ޵޷[[ov޵&ϐ=s^kgw{ֻߴ[?hpsoaWWûcI5l^9 r4xڛptNNodû[^qc;α|k<<M6?wzfpL_8apsvnUtn\H΄ܯn(^O]/R_ s0~=o߾gFGgfH|dw= ΀«S?5yb>宐|&W[_O>9 ӑ/ %j>Axpbɏ-FOcLռ@9}~k2-d yuA3?|j;ΨcۡkR71OG3B_pxpOpu6kCT׌?!t ?l߅^&tbpoP 鵬4R{^ԏN|Ð!awg _s'.^c' ϊG[d6POϷ7蜁)1-tz_wOrk~j5(]~Ig/!FaJ314f^'F.'=aK`|[''[us#@ w%?<H/gZ7x1'\q;$uipкOO2а "x":}D&&nN/w+(7(4BD _L,SXӌk2ʢFA۲kSlnߍ;_E%77ka̗WzX- JRsnUX@RA C0˒ 46;''p4QԢ6s\pk3ym6Yul,۷$U6]}PجuW`} P\fZx!oS*xBpo.H (Ί0`Xrf `(c>zu\!D!"VEf#'v#i?(9ʥr`&48fX KB.(Ǖ(<A^8,U1.b$9VfRĉRy? !*W2yp'$3x-s 0_,SŌ @ю^R֜cĆE꽢IkE$9c4𠝈8)E 9ǂu5Bt`O5B\*dArd~eҊČSv R<!ƀ dim *m9IqxAwVx#5B[ FÔDbp"gU!%PƄ J[=tpTxYVHk81HDrxxXxx aY).KdpJn9^W;`9O*fh*fhVv4{)>[F5)%؏' љ==0[oT̈́HEf0?k`OL,0IQc! $g,1e)(^K),GT vک3SP^~9P] r 2^ý%)+iH9on*Kn*Kb2IMu?)\C y 'UUTU%KUUTU%+VU)zuH#y5J;]K+R=eK(+:]U%uj(/6ƞ#H .o.6֝s@jf "fMi% kgT0zm9T)|NLi@ʫPEܴ+OU̟aYvcL2|b̍4T'ݴÇO~`OrmxW|,zA ]ve('` SQ W:Fvtj`K+1'5L%2f- {/ U`()*#s`m>~޴\rqT ![Ir,w*fSvGR 0&D qF DXioZuhk㧝U. ))l`:.h KE1ZQrhe^`'Fs% Xy{Ev."FFˬө/Qڊ;;b&4k(. =J/Rv.KcR0KT0KbRI05K/|%Y.>R2U3"TVq`C;o3X7)!u\,˨אkY`[yw܏Ԕ'!@h%?nkO\iκPtK RH]S@0)vK%+M;EVoWXl,_kn0|zPɵz97:%& $Qv!{j_^L{rtfF"KDXVoU 9⟑Ԃhl[C"q4<]>1F:h8Z/ }Ly*\3"׮G躪Bu*Rh*,Ģ~XIjH%̆TyP{g ω^e˾ QDz,2*KȾ 3ļg)KP}K9vWua3HBLj!z`i"^c pDb^9ҠժuAx+-+TB[-/h7\>=9eT]uVKu"Yp${җ ̟l~LX)u3] \]Cc% \(T!60yMT^`$zظN/Χ̛zIr|ƴ勢_|Kh67.mߓTeK蟎NNG/7_t>lF'/ jV6c#w1{Gq7H7t\Ϲcfe cVS9/F4|7WC%n%ٻ8W>d6}6V m<F_%,)r%EQCrf!n]/ql.Vη%1Bc,Qu5mBz=G#_ؿ'WXb}S[+hR\bLL )ǏK[W9H[ ڀ]CW{d~q?CCTm\ZE |Q2;ǯn ( j_{ϰ /xXu U8WtREwŦ)ȏ>3sJ,P/hϏF6O"?ϫ"afޙx5WZ-o?9{EJoONV{87ՙԾifJ0G߶^:'9yjW%ݫ友të<=ly_MHh#xne}&(J@-\SԠa5<a`sw;YuVޭj]<>m 򲜟}jݤOUW4O%l.anKKE&嵹y׽ z 5Lw0ΗOG#u{m[[ooZ#GRF}W=5Z{?n~eC{̭޻We岷˜ՌN[asRjCԭ 20u@[豝mvWνom$C̽ \0dW/y!孨nӈ &OH9j\-vU~kr:NdJ93 s"7f;ܙw&GpJCs:9ϽV+ ajgKAb͘g##j}kWo0X{f µ۩vﶵmkXBrm x,ڬQfNG4:kQ&κયͭր)BPZ? z;vf~;oOt:ۙv3}<~Pz #P@2cA*]%uPbTK1)vEpGgVȃ{{#8k~2 慁m'AZ4xyvz6HɄ2) Ud2*_tVŘmk E~)xTvRy"kf3}N^0Sܙw3}#Qnag7#38}?!^74.[#/|e@UaԐT-P#ƆTsEQ-U.Cc)43s<'S#:sg>3yēI/FBR.i^9⻶'hٳQRg]ݺj=9y[ooH+ر: H5++/߬-e=hؽW*xiYL*geÈ ClsշV&*Kb;c XAg /60)TiܕC j\Ws(T4Po[s(ײUrWu֫\=?:|TUo/ ](HŤaQ8LP@ZՄĹ@mIaԠc=뺋]!sv_5x7'!i:o6ddi̺buǵBWmƳ5!j)kjL c,[YFSRu@%״Z_OE9MX]QmCΏq|>+aNZiGPH/̋><~hmyWc|+A=VS~sthNZU|{xP׆jQ5Z+>W =<8 v|z|72!ep-P/ٵTjR ."֢kI! ՏǼVmߘ[<-;I=KBw|zL9UW;m`Gl:I%wW c6K1LCάDHb5!q$(%%RʻwRbV^*f)FX!Q)vKҚ0QW"\B⣅i|rД諎-)ݟ$vq$wZ$DYo olTLpMAY/:Z]ij!pd8^ pNֹ̈ Gmp_X'f3܍n{`m }c̓C@f w__ b*Ip`Uu9d%,a/@[| %*{Ofz4&ʅTQœΕ.g!Z5y?}F nGPA[d(WͫQzZ@81ބ"4 a: m ^^yLkqRGx;8 5I[Iɟ\efu 1`$ \uX-tu鼦.[Ja?$ oVD[LAw%N&*S5:wYb+a H>mb4I=čbng.ݛ1!Μmj0DIb91m$gxLV0yG1g{aߔQc !#9\]r5Bޥ~J-׍W"h+=m3)3@'{lۤC{@qyOTݗ݈J^VsO9փnt7&o>iM)bX./Oů̱@& iy^u8NR-q !&\ܑF͒ uH!S(bdMBL4̴-\jo)Q}Y .o w'7zsF κ:^w\7Sb~*:ljhݐ9Q -UmzTYPvNKrSSx˓v ~13 Cнo 71P t @uP|DH%fy+.dBgg)r+^YQgR*g!c*UmF֪*z{i6h j*Ԗ&J2*PLJgS2ʨ62έ V΁r*9;B$rb Ap_=H‡gU I"<ȑeItUUOYPmlng_lq~ CD`s':P 8׵\S\Tk 09zE g)NUa\49x9v۵v/,H:jvK2lbjIPpro G+0C#6"4 Ti Ǔ,laBdSMaQr>qDꀆH1$9q<_3v1rIZHMx H;4 k 5=쥵v'99L0|,)7Ӱ}&LQTpHRAV)(jTp&.FZ$!gwի/b!pKrf$ &=ܿM??aB{s}:I*pFFEF DĴL]A0N/Ih@(i?L!RE l  ghgovZBVuzOsZGi߀@6[VK@H1\ йb(% s_xbFGRP+FNVJ*" xCXW;tuAڱ=YBVQVe>P]0mU'*%$E`bi#䀃ԯB&8bw86*5ki=+9bkT疺-")|%:#$2;-kn6D 0)VaS(WM0D윱#~&Vcr۪1-ڭlIXtn_MP@&D1I4YFQ>ԓD PO|UfxsLx[1*:`qa-J+.V`)V6gQ+1YVNBӐ5'YnW⍠ʈZpy*̸fUcrCM3"6${IVk}5 ( '$Yȝ^LxYPu#HPiR85M)@AD hUc@8[5&ak{@|{Jh˻]8& XJCC bƊP@D.}b%5)+N @n#0޵'a1$WS"9kI H#)KNѬqJ{"ݓSA':wԬO5]ǨZ5I`?Iֹw,'I BhȋJ6O]. 5,A$Hu!c^I,)XN"16RbdŽ ,7#q)h֜IڜGZ`95{5eD b[Af\D(HQP$0'UH>%Jk*[L 0ldpp{8&hDIKm_W5 LۗұF>UcDONs3QIND:$zhMHjIۂ3i1){t ^)*bo !V5횵Uc%1PGt U2} e6=K7XV\ѲzӽdD*Dy>5J+i,dmՐۀaN='T}76e<=u6 ){#]z 4S_tš)IofknfןZ^؅- •^&-z7\U9ҍd,N`ipr_ìi1.ގG/?MUWd~lR+. !,صbQrÛ$fn>{uMeήgsƜ/" Cw]8Qo&1ͧ /\|b>dqɚhے< Y*@QɲC)x%\'|{V>P7S>Tᭁ/cQ(: >PX"!oK:_\؋k 2Y g%Zbj{ 9ý|qS*Z(i?jB ^&6|C>l<*cVi#!{{sxh8li1Jr/}8)r=,釳?ߏ#b(t029z=dZp,gEub9ix*#a|zlV9Z@? =fGf/jcWh&=\e93HN}>3Eze,9[R-Z_ oGgo>z쿲 Ïf -鉤u Wr՟]8$UJHQ2ߖ+. "AybO(@@ᯟa8xxeģȷBg:2dZ yVo62!jq66/-#7NfnZUqg}8{ yE:@&aFux{ }9|Tԕ};xX!p6Q&)P):b.1O Rp7c7 mS7 e`Eq{x;>B2kqPlP"vb8꼓ZFx(;8w+cASQf SYk"7/vr3_9|F[J[ThX ,E%BF;a E<$,',,RS!``DbGKzKG6fi&l#Jgɫɝo0֣*~<+jp~MjLpm>t{n|{PpxNP#A8xiDً(S̮ET]z `=fspBiʒ`N@ghu8H! SVxKFWgn]yP:]D=25F:ZW!9D;`J]̗qh7yt[G2w{(BNH'ݺZw;B>sƔܨlvs &:D.Oqh ώFݹ{79GB-峷B 4ׄim~|}yy`3hz릿L "s7< _7Zfוz<!P7,rWuL\cu4 ύ^u-Bs@a)6?-Z ("%$,"iO®H@WU΃"KsaȆu'6%åZ0%/gblJSHKa7'1Dt2?e(} c9P/t𺬼-qYzgg0.3tC6ٻ=8N1V1bf;M+z"FIlɬ40'g-v.q̦#O͎mMӀaO43!+naF"{IZ:4H"$Hй&H:.YfD9xB8D9K{I wPWljCUIUq40 SR=YECHN9.߿0I>q: 5N?+0N8j3NX'q<2 $ʤ NZ/|bQ"\yW"{ieh]t-[饈X&Zήvr0t~xkK‰xԈ~VMO1Md,S_h6t7brec8Xy|4L9;**^b[٩+SWhN]mکf I;OAp^U9D)xKB$YxFY{u}P[[WZCqیlԹH@Ј{ym̫a^W0fZa]B7wWOS26)JJ t&4GywO](Аj8ފD"/ RD89]ͬgZF_QYrDvz#{l(5$Hޘ@  \fJ~_!+3+? |+MU֢O5 kL(RMb%d0TcRW#@ c阓Ruȋ`/+Gr5'uIRnNd^x[󛕾g/s'Y297ܖ\f%^ݞ_Iյl!tj}U[;%zc2B3gūZ#&:Րc0v2Bw{ 1^իmՎU*Kh g0hکD7T"p6$a:l)[l#bN)D4kMSUDYl.rٶ}@xd?[t hTĖBE*yׇ,RґYy4w0s2֨ӇHVR'aXJo܋ciQ|ŅОJYeFKw;*xfyϭXрx;Y .}G0' Y'½ W"]ѵ O=4$vFw79SպG;!*;@0Y _28IL'Ĝ"w rĂ32,,-Rs掠VA[M)`tyJ4(rdQl"0ǹ9+Dʵ0+'mmR:PҞTW!UNAVw E3TIr^Rj)I -RN6EzJ(u) u"o>%^ !ZB{'uu16չ}/ccw2Q>[\=Yҡ%(XCOHY61*}A~1w1FQU k=?-M&Gdf/ߕ_ͧ7=CX4}{;F?yg3lQ(L{oWB].ppWzrĸ.QГGOD$]m6޴Ox L +p^\N F^EőDxR?;G܁IQiIcϧFӤ ^ն;O]N wkGLSt :39r"*ڋ)ӠBD/BBisWjzTHs8s i dٸ] ڎRi>Cv<s K$w.AdI$FQ; ZP%tpm [:@ z{zX̊,N4*GS`1b(I ɭ;}gDrl$p sT"/b\&E2cR#-TQLmJh^N A3/T*@4KbQ( iJZ-T^ zR6dSk!)f*eFUE= Ѫ̻)ǗƷb݉W]W?Me{3\=O墥׫41J?(m_Tץ".uUo>>D&r~$IKԊH5")Sl2U0:SҸ@ {mW?>>ؘo%޿Esfo]p;UAJPX_?.Ǭ LșqaYAi@q̠:܏ٵV*bR]vtjqϩ4deWLy>zZHU   kLVjċgO+/ǾU'uOC*xEШ_cJA{AT5_x !D1/}܋i{4i8d, ;ol'PT%cȊHA,CFLʘĴ@&U ۋs\_12'(JQt9?aoq?k6q5=Vj8͠[l/^@f4֢O(\Qǂz=lX7[*F)ya]^`}Wg}7X0xf{W8}6͌Y:Q3j%KP4TzV8XʨzM5uǮ=v]&"4؂Q*v?ƒMbbC) _R X Py54s| >_lF ${;O5g5"9UobSZJb4k%2 rCC i*s& jc!r&H9 ؉e"c"|]*V 'Ԡ\jJ u,<&y"3L'YqO MHX'\5^,\.yggZiXcYcDT`=̑i諓h Ogpt dѬ@0Τ| rg$X1P^͝fCk9*ٸnTZBrs8C~+hltX0$9ZKVE6ЎdvG%9yE%ƒ'>7Zi̋?vF4O[>Ttì#lIeQƚtSdU%%RCd ;@I&ۧJfͪNbNo>X%G؂slD)H4RL,Ub6&R; X<-1Tw9|PWP V"BdV ljHYBlZR6CVB)p&cڝ C@OGET 6<['&wV6]ר6ik[2/ 8^Jr[v>揕<Ü/F'vz Ʃ&3?Z4^gԡI8g1bTjʽʲ8e!XmDM~wÊ,,4`,"vgg@))>*;'Oȉ,nwuӮ˝v]NF6.lیsscs~=*AfS*˭L8_ܙƿanׅ:F]?3{Z}B5فo=QQŢ6O֦6_6o|٘N |u 섭~ʒua%K@*yWNmdEÍ=iI( Qqi6j7[*1 k񂉧ٟɏDVq' R"6dӨaꨚw@,Q_IRc$)~ߗ_7_yHvu ^DH VR.2hosoEEFa_M"AA.dY@!)A\˹kEOZrGڀf0QbISQIh 5P!grzј/JNy 'Aɧǡ-z^:7@ HB)]u$>u$:h& Z l%F>!؇ +L,!VDXww%Pf<@ZI }FE{Dr,m}D2tB }b jKbRi4P%.W=@p%a$.JXnO4r܀x#˖Ӵi}GD ]pj[\:H4 }-)&XHK|hy^F>łzrV"i[Y:K㓓>iF.am۞%I)B͋]"XZAU=$E;8?]p6T{Bu3N#@Pw/a!__MpvdD2!Ovc/]{/QZ_X#+rTJq[Fe,.- qczݯ> 4Gs[FP %HF)̖my4swb8YM|=VY$/[}_ﻏ ay=kI8ȁwA5&Éx]qޯD^41 B '{{k5dUٺ1A! - dd95B Ĵ5+l6 ; c{0C_Bh]-p).ş?-Mًw7* ~߼M ?>^U@azJ``X ߻W}wih{Ƌ7OTd:SG{c^T;ۚ5_2щ!<4A>>"l)VTq FDI>Mm+4fȴu5>ekIz}^n۴u}c[i &RU^;jFI)y)G*3yĨ!")w0u2ֻOfZ :z1*z1NF,]*%Pu4T)7J xsҼ1{Ӑi>snr/֘!BI$ мt82qضm cO%^ _~ڱ$>Lo 9WS$=~v};\)wkķ—(s&{+hly*6I[$GUO o4 s1jqjȲ x|6_b:X~)%]8m=RZңJ&.zyM`[P"`4-Vɉ]<}Y09xz([nNZם#%3chi{Wo凇WC_v}_p:u?z1w_nʯ{^ۇ?=w5mZFIeo6Yol nðERnUIm%v9ӴD>uDY㇆9IJ޸?Jf02amk·COYEWfU]w=0< T~`ѡ{C4 ϫND ]G #գvs;<)pBvgSZo߁I89?qqwgzxUٿ_~sWhwC7n{~/_;_gG~:WWH N~rbFWGy7Ϯ#//Fwt'|;E ^UW lShG^r7ӝENN-ntqTmáeÿ`W,?/ܴBsZlXe3 5VY S妝M 7%\9Ҁ ߯|6!~ݱ ; m_PTY!{CB\bxtWKt;ڹyU;!#8ԅH~Q9si:k7OP t!u;5~xyresDu#;!Ƽg7N\=|C${V=ڷև?0[uQ;x;~MΎ~^A̟vǟ{h슸?L8i;Tb?CP V G=0="8ra?}:cO}} <~!: bc+V7h|PQX=Fr#%r+QƛouBDΦZ :  $N41C<)O7bLl7Z<71Y;qfk)@>pQL}BgpPf;8*S:nL8ߣ?n`(8!o5Ess Em1+i>~7t8˼ϸe& zX'~/.- VLK ]jRC$];+3 ]jRC%:4LI(fR$XpJ)ӟ4-hRIuP{>\˂% MCm Q%e ts+ 4CF+XI4r$龞͈РDf팻7[.wq*ζ34fg"BP_is6v\O8ǹ+EδWxW–w/TΟ ?=W7h@@p5H_%DSa n.uD@L|7"(aL#x~7:pˏ]җ#Ьpo!K_6V\K &Л7*\{rkĩwmDEq}- =Yi;uN ,Vwp{"T+$Ҝ7DzDzhY-F u:mC5!z0IcU2m 5Dq'阙ljP.T,8}7 mT>g ͻLϙrFXrrUn .LP5ԭn uk[CLId2PF݊6 72M1FVLp,c1S[Ȕ pcuh\{>\B'T_GnyɵX"y^MHQ* V=w>,(tv_~RJQXmճ{z+ u+>;{WSWqWO^p-`r-fPyi[c.'犭I*'>>^? E52>.]dgrSEpA&0цD]?z}9p /a}:U+ZcbzCIJd Q,6ԃoE16spOһ bGIQ:\E}l3%8t+B&4 ~ZN .6y}=pp pą@!aߠ'!۠7((_ D&c;|kI=YmR+h̰±,.!T}&0?!- Z*WohDorf8 &0Ʌm{駈U;2O6iႁU$n *)^BsoQנ=343qoյ$%mF-13F$Nm)382<GJ/~R)+Y25,z|EEjТ#0mMT7jJUn4I$4 #$@`hSbtk@Jw1"q9 |_/E~G2^tkA`/?9+QX1,8Gz}{m^Ԓ/0׫4ZI[Q(廵>0">r>ZB@(FB}شkW:QJQb)JRDUG_N_nE&=㮝}L~LM!{x'Лv-ĿoO}^$Ƿ1ٗNJ&ZE岺U~u^!^禊R_UBlL<^}iOWOWki>E&bnHskŃ[xIl}]pÃ-ᨠ ),T>N}Uj^oZ^fv[IX8yKG oՌijcLTS47Ν$~bQ֍£~/idh(q_=/0ͳR9y/9m FN?z$4 RExe} iDV-.6\,]je=Ek[(Cb1ߢ\&3CSӍ & |g!WΣNmL9cA;zC\Z@p>I f$ Lj$K҄[F)P3S KQ՘!!ಱFn' e4G$^^T]WJ}Tt άp% Rur*R$1K3RqIJZlTb(ȤDLb-0rUe԰4&W؂g׵f3)4NHib8YF52Qlu&9N M 8Swz0L) ΰqB<-FXbL(QO2z13x\dgN!fQ(*La4҈pZIP*V19a?15ͩ2.%eƛ&NɥR[*h.Na$bX,(HB0)m!\$$h$)w>C"iN`Psko@W߽$z+*ۺ3"*LDrKUO <<@--U)_t맒 J6˹ %DlFjZ2k ;]7f45©"F&fcE ߴNH s-DT.Dp64Z(Zsh4)mB2Z㨒搳PN6r 1H0Dݒ%R R e1f$o V ΁dK_ƽ7ϥO0es n $pj>Dxїˁd3.`5Db*z><8ۻZY0pe7MQ)U:)OzswuN u>E(\)"~\KN‚\kcr;qW߹񠣍1n8.'{rwmݽ98/#,x&s4L }Z2gE`{;̢_0wˤgYeb4}R@bH3Bm|KT[X/N/zd \ ߺg"/LTK1}4JK_4k6\_ &i&-kjMj QA`ϛa ֩2XavlvhbR=sStͺol"x$n1Ÿ!o7Wa~.~&-]\>( f|/ZN2O|I; ~kyLZ-\%c I9<@ФE/ a$fZj{} SY0 aX. ^x![LdJ!HҜxPSJլPN,PxaApI?x9pI Yd DS <.%_r;1 ,+y )*64)Z7ІFFbpbYV`)1Zieh"% G{B{)* UeL92G!#0r *r&ee*VDr-rHƘ (+9L#1XE2#p.ChSrcz-㡗d6n0+);,-iCWpC a嵤ɹ&%pL-/B*Qze&/ fW»mQC3\+kIK5drPW7)rBr#'guQ\ZH@AKSL bK#|=h2<+K!*y6Glv>1!:Iqx+Ĥ\qUF b"=[ 'BpVA'-Z}%CRsJD 虿'%3H#FbZu*ӄ,찊0'V }U LW*e}(IBC,*\ml#/M&*,CౙUTn_( xVhnQCڜfFJӵEMf-/ l5n65ȔK)PFFš-RaLdEm EK;$u.j68M1ʚF{4r):չ\H̸| F h@`YR|TNXvob*Wܵu {vЕ\(yi!jѿ Ny\ԙO/nM&Pɻ.-G#.5v';r VַT5T#Z2t,5-.Uy{ထ}r*H(![3H4F6LuQeF7>rE_Kݭ4ׇǓk{zC#=K^ZGlǭ.mH3fld26Ya?k וVZhbu?LIEvwgTOx8{3oD !;dwӞWf,"@v$WNy0jgOEŚ {8&mx~q$uL\M5vwu^prҥ#z'ŹwʩnN$W|f$|'ކ#J0NN&Xs>~Luͣ$ER02}(!JJ+'.J= wSuY|U|ͫą4nj#fڰ[TUeh5U_j~f{[?f2Lwk1֧ygb7 >ZHLʀ>{Ek94DШUketLS!tz+Ɩs ֌~(h ]j]hd/vJ -[t.Ib-ΖCGY"Gpk**+Z^7bi"SX3Lqo-!@A6U|b؀LwBԪ$L"Loc+7zkG.y9P)\g!$DէḃouW< Z -gD*&ё ZVO/k (Y,x)&d`O2yFpYD Ru%%yU bۤ9=EYu /(/ ƃb͎g?7Gcvob}%a"Xzr8&{j7Lٕ;w%eBޮ8\Y`Մ!|Ϊro+3,T23VU2@bo^>UZV`_>'4{Ӿl d^Sqv +nƽ, Wdoye@_$WO?W/ؠ]#7(.hʵm'<=ĥi 4m~=oYi2V럺dG9ByO"7!Skey|9},ŷIqiy7y#9-H5wv\濦p&ٴlۂPȼSLMOsƢ[B' jf}zei^ C5|qZmVJW;Q>c:鏧4Wm V2nin4m4JO;a$GlST#M*ǜ,*K[qꬻmom n̸'_t12cC3rjF·?XO߼l^FC+ !qAH'?T(ˍ>gz,l-j {1G u@׽@k=x:ӥry/G(oѣ9{\=G cϽ./[]T_&۳ПmKp+9RBˎuM򬚒t@)kQ(Y|eu?#hjt9kAd4Pe m2?ogfՊ\Š?r܊: W/%sW+ނ^#9|ȵ@݈**ВX93JֹJ|V^xdtVB0[,KN*W.kAXXWp,Vx4<\ 5A]Kw CU41έf16/ݶ(pJ ӏ7Z[ٽwlq Cm^NenTְ2I)}e _%r9DIުQ) ;a[.׸5uEkj7ꁶZ'Yo* &78EU 8`cstʒZ{ Y6ΗV@Vզ;Nnx5O4Y<qNY&V ߍyp;^/׿ſO6LP@籾8}aPA|b9t8ZKv7`=&-Y(ZkA͝v&0Oɫ?,V$ ~|roɫDj2,4cӹ/2/FR64+nrB*>!y0}T| |feg|rq@,pvI|Y!/|F)Fovפ,tA҉t;+E*Y)зthtb%49c/Uޝ&OWL=%];Qk~zp ؇+#*1㽲.. W"p(K2N(9XDù"vyzPW"-T/|>??WV}9zc ]ݡ[dtE&J}1Ps`O[)Y I|xeM h0rcb|&p?Ig3̈́u7D)o52V:Θs.)c 8hg(ޮ4JF:l6]kuufˮ؁'۲N|J=}qkui=XPprxu`jsd+,Ҩ5Ti-bpV"0@5ƖI0M_#q*1k%;Y4'mǬ z 3Ӫ{i u%۹ _!|/%aHLҽϏ7IX&)u hoc7svÙye&i5Osrå)H*8u Tm1B^'&E%n6N>հԁTK9dj^ hVi_0f!ESc ՖT 6#3ԎyFd;.G* J nj Je ǞJ8u\LeI <Q)Ds2{j cfޗ[uB8vN^4%hGɹc81ϵ)q0L[.qA_2gQ\nNek{OJ+W½XFFZf gx.21h\9A*K0eNqΑu,[zͽH({UkHpL1$>T1 cz0Ә[a-dfgNEBq%*I}`G m. LZ=ĺދ pOad?|jZtb DA3iKler2aDeKq$0,=SX@ qm(9 Ni7"BE{\uƄ쐈f_Q9&"@"Oj40EC6a8a:j#Ȕ*P޹9&&F?רj\ yhHKe}wwV9XL8W [ :K1Q΁@%)*pj3e ]>XO:쯏/{5/ lFPmT-R79T>Iou7%^M`}5%pT3&h8D:.džj #3sG@5p]ح!3E>z`h*I(J!T=^܊ V`aA<1ƘYm &nӏPuiM ~v$R_ݴ4y?  ZC/R>vEHs9'a+Yw?aCo-] Dz6Fhpu'Nn RǷgADR%Voyߊ oe4{>&s;0;k16S :NՎuD*AE'#}W=笵\SRjcSc**橡"EP"πHq,74n6!)%ljGT1ųzx$ SqT eųcK,H-bS4]ڐ~ ਇֻPovx[yv:Pa6O~ⵒK&[-n=DlH3РՖ&/(XxMF{/'tV7]JF.>*'U9v@]O>߷o!obj|vMN_b ܁[2ehf9m:߳L3MMХ5k ڌлh ݐ )'+0F7@>m|)(#v~n f\iA`p>2a|!H1&K?f6ay=wokV/fn*IF+8;o6*wշVH.` /.b,*j_/e\']L|=s]g2?L,C|N+"1|UI!f_hPXQL"HW2pJ0d-De9uu+I$bOnIrӮMKUHk^હp ;lƴkvN>NJU.S>(> {xK&3fQكLJ⪤k mqH$?+uڂIγqHM"`7&"pw1^nIq3~?~"XLs ןK7oGic:rPv y~g'tA Z9ƶ/珳I1UHOʅ3)ʿ7 E5~jLן./ 5fmÕnljq&6g%ʐ4!V,ͬ9~K`cvY J4|@ }IyJ#޳殠+3hFdgSķgY{+<] F]#r4ycplF]狉\ 5cr|)Y`OHRBtZ,Z#P-p$" (@BaS,5eR`oSZaB)JH#?h9F_X~gEy=cg5kZ;6gK B/qాΔ8D( %X[g tϗA3DF:C,+Q%ETvl,xq[a1f/%c6Hk skv3ֵ9GwZ\sT'ҵ?;@B KӀ똣1_X <l!xl)//0:`#t7PGB=_ 8O=j'p%Tso{^K*Nc~A]bT{ 蒒Yo"=9|K#ڷ:H qx(H@[8`JN%'}PX}d"Òrho,*?JR_O#1+xQ'u!kB7֞0mfjFm6NO@ = zxW;U'QinNʒHjz}hf5ZXډZѫp=8s' b P\rs%.L@6?=8dsŕŒ!7h'[ oD,0l+'n5sjqb|,$x, ~K$/4I}2GҿW{Tez)ƣ0W&p̈́zc g utq}wɇ֛X#wJ)$ J J;~?[G_IՏ=O)Ƨ4frSM3)!8*g b,N,lrN sMD,Ҥlϻ벣͸M&]On4m1Zk~V} ;|G~"DKI:e=GɥpYwV!Q!cZo;Yx V+y$|{{IK;KƷ7Y:0wn.ͦ@ &F(L˜aZ%s4u3a:_̾&EYsJvՕJ\+DQ(V scƤ)Fp鈅 +gN-.+r4 bSW"U9&~xt""DX &'|ʋ $/-H>%Rt1=l!ͥTJ?zk ׃Gnޫ 6c,ԟ it#fouN$ (a*J6zxW`=tW9hI1ÅZ,ӃLFWߖڝ(n0Ttx_ UKk+&h: CXC+,1Q`N EO6eqQ\'ӬQ+ڃҷyhױ_װ_Zi>3:۾L$/qn8VA)fZGxH߀;seo 8ASs+Hn()Z,RY2D򰢮32}?vEu"h \N-L77<<$J(w_lȄo߼9~/@I3o yB*^c>yM(!6B!{tS`w`{-l4#Ðbf zoo}?ExD`w|}@9LGړ r$`l@t9V9g**T֠l} ?S&LU3ݝJ9`տ9P.}3|WsO~`0Y~qu">7Iy#J .SzfJvz=YBd`@8\T2uIO`fV!C6 `Kp`M>;=>Bo_ kYIl~ 6Md3N5A wYFyT Jٯiy .Othe&.KvQpK6U f?bIfţP6 YP>e=TJIZYGI"ڕ6__1JT݄OC}rGaS_11Pt߿Q$\Za1I0 4!>jqf KP8",sPٽs!0u8C(H0\憠3#39~" S-n[VI5ŻCE"ػ6ndWX|ٓR*Jݲ+Nv_ml%ErTi I HC DZEp|htc_k܀8rɎ⭱ĖYȄ& # TġviEbGsu{ rwU:g*b;;rJUǧll0a綄8fD]PqRqړ_bn[0 1 'n6j2Z0bd5 ʘ BձzBJǼWtk> Jy35}MINQ>LVC Q?\YinF)*tң|,r³2d 7ĞGy}/.AXVI.p\EV^4]GgTp;xnxT՜E‚xn/"f [6Nk EBzty%F MQ p&~'s: 5_ /%PܺeelolcŧlARJ&/GIiu`̉ʕHF<+r D ; JYqLA|dVJ[d5%F- BWr ( 9g ;>XVs%"1+mXuWUt C-},/<#Q 1S?d!Qw<ǫeT4t/3??\ٖ]pA՚hӓ׫t#XU@]&iOj蛫ql6Pm4{hY P%*ǽlsBzhBTҁ ihmZޅ'Vh`[8I[wq-;_N]P:^~'0gPq}UsE'6_x eతþtw .c߼E7Cg¡OXsܩ>->Sz͉yR pDs?f\9͒SR-:x^aǪ20ScR8WHb0F8XSbPXY*BF*JlyJaj.U͢+uվ}Ϸ:0f{P* <(uuA{Jw^H](*D#YaeX&4C7룦}rbCRGNY(S'PF Ԇ|*ZSJ &?nAXR rT&dI[Ds[:őRTpcۯ6<L4O=X`e Ts~u^LRqڇȇkՏ^=,3I_{ICfR qQm _7E0SiP6HQ1KD#ء0H.k*$ub ðeP^ k &2'P*h3*/}9\ۨ<ǁ_%ɬnY0FMCmnی W77Ce1\ײ kyY(Y(֟2EGƾaVaPSџr0BY-B&/op-;o޺y'D%It"[ֽoqJu.l[np7%r'QNGi>ېT]1,6hLfSH$9 ${u :$${u A}AU`:8)uᚠEWQM-kz.*4$xÖsE(G(HlI1n#RbGU_f{1ϮhrZm:Kni %Y:<^)N$qTFm,bCFTڭ2a.Y[nr,]b Ot1BWyf*cF{'5y T{Aq\i N^ВL|@>l&ӑd[U+]vf8$ /](RNk0J)K.`L3Bo.7* #΃Yd 1y}L ]|uep@pv6~}uqn_-~A*j>̕Ҵ%z-B h & {+5e_@Syd*n܂^` " Et<5,'i&(U5!EX{\2C1yg5+ApBUNG;g.o }1؄B/.&c-Iki+΅}8oﯮ}঱0F*=<PG'XrxK~e $m.aHpXQsy{%8E\ 969p Њ8y@1G؜_r-ۅZ#_AДМzeL7{lc? Ⅼ}34!AH׎Xyo"j]>CAvCfC} w WSpZkcLH lcLk=eh:l48KR Rr^r9Ix'fb+sCsm0pL4Ij67Y}QgWP"2H$h{#;± j|)w̑BQ~+ ]O\B6/>Wch niVEk 9S`:1?{y9x@X/3[C^Y 95/';RuH_dn+l'Ln%2٢xZ,vtr?6gf=/1jc%,߳f✠4VGZ_KO NdXqP ؤ\FB\ D㊙BZ 6Mi=ԚqOg)M%6he:6btI|&hݝQUa4:;$5\gGh\s? LH9QRBފeQ n-@ޘQ 76)ުP~!{9o@&K6[ֆ${@J,c7^hUOq- 's4;ĕ+ՂG7%LB?8IHC470̲|Υ*L2ˤJ/Z6zi=՚+P"R-' 3 ॎI`pe;Ԫlg"+W}W⻛`p_{o%- O/߼4Dr\1Lߌ2yτkp+/3 W ~ SoSgcQC.eog'Fƛoy0oI!J#g3Zf#,/k!_yqFTL=nX]YR_O"0#3)90-1:ܩw6\lUN-.Vw`г}$S!1qz H 3i9\sEv5v2yQA Ύ|\{lp ;e1Rꑙ4rRUי|PǴ6-BKMsBc{=?1Z{zh>Z6q^CL`&!o6zڟjCd9 Nݙ5Ps.TX+tJbtBT)K"$WF>G㥔 AF,oȦJmgfrJȮ@+7\edžDL9fњϨxEk M|t狟?jﮯܻ-^xOCP Z`HS9Uiz΋5:q+eEV)*fŋ1D8}3)QV9>Nw _1><˒C$CahNQbIkJB'\{-B **WJ, ~> +׵9#mo-$6 Y-c0?*; xcaP?-<Z1Ä:RiiR+@ğÞs3e%S.4<]?~8K=MQ?==?=~|,.>yRLVIT▚:d]ӓnkI ?; }z䇗_/PlBzWxz(7Ј%рZIh?߶NQ߿n88]cZЯn_pSmSٜbŘ y2K̤UA<^^{)sєଔPpSe鼂b^8+hj/z,3j8} *ͽ_tǨ7G˞mOy~w._ۇŚ wb8Ku 1P5G}?l~[ք?|iʟVO^L:=d@ kT'jRAW].:⤙pDSi0 GNuP&zH3D-t|wڑ3WN>|3Iv 9r,ȩb?'mJW_F^y?/џ*iB%>r!єkR}*v>GFKTpY%28HVo;P*mo,ʔs aLR^&zteҭUZe0S' ))J 6jA0JEgrہZr˲$j(26O5/!j[pJ:'",QX8L ،dl[ rG(`#>Y1r8.A6SR*:\E%:*A eƔRKsAq5m SI|NFZpSJA H0ϗf4cgAH%Z TM(]Z.h*DY o d+ |k A&}^Qp;΂qT;ť.,g TJvi QPGYJIBP[b.l K@(FZ"x.BO \m^6s<_[qp,svN͸5Wxz澲|!>-AԪ9eqfF4<9cR]_neo?1TAX*{/L׏gNlOC[#aKLxß6k8\ۇ^lһ^sr5h.QB\ %M~wymo=fl}zRIX)KtԠpfD*qW6qc(71TREIc!xr*͓6DZ#7 ıPPUG4থP]l &gjKޭ/UIMsiS跽-aRHn4LeT$UGSBܪ@%FTB2X׼sPrQW؞Rd %h6 >-ܼUfᾺK* ,{m᧴~ҵ;WTP٠ȁFVm zqKMJu.A&Sвa('AdIIP -\Mk^]u5K2yߘْ!nNJIFz n6` cuYDx4 ǜ)xJrEr[y$~qk`+[`0]'aH<^ $Hկ:r.Eo֒~;mvj=VGyޭ1T]rWLU8?)3+n8D NK* ^{w,1+̛pQB.toA9JX[Q7FXO4{`lyt6ɻ\M6BCG^ToF67vrଫϞ9ϴ!V<[>s-?πjF0iɃC$\[Mեp7xٷ7v2Q @ P5%u'@ 0i:Ե+8w'LI^z@`a0##x|O=C+lWҜϟ~PZ97`FBlu TWjjwAɖN_Z?_4?o; t_ެ1Myc /qbǑLh *:`W4~m(xSIiƛCRo H[\-xal#H%]DF͵ss%R SFF)$q #ΠLhh*HJꕔAVIvƖ"ZQ2.YlUX;MŤc|G,^lWg5m>^G(fVY#a!8Zeqic`V*ɴ}.jr[Z'Y|S]zj>CN~lN-$ Be]*J9 av DpIBGG(7SjZK?yATSm< OthMĘz3E'cvFtZTnc[L@(HSӮ`RC|lEPwٮ6_2g\J%Z%77[ e$I /Ym><(nw" u Zu[AމhԬ sRcԪX˲h9L۹,sSK%.:"ɱASM,їemt4V "XI@J9>!7V4FR=`JyxvA4gNkdXO+Q"bNjZfFJv&T$hYKS^;%(\-C󂊍˹0TuIK?OLŃpU;D8@?+n1f>[ns<RX'۳6u' :JBEPzU 8>T!JAjvw~+qt[K^Z d R*BAYYva641Po* H=s:- &@}y ֣ɄF sx%V0F! &::JN%-j̀5:C!-3 ԚYKlĵ,IR?{WFAIyEn̠-:LRv/fFU:٢!زT8|Ŀ9P<σfh,}J-²Ly C+NN[ -lRv F߼jo 5:#ؚ#2dT4ǏiSìkȪUe+%SbrJm*m[|@g' <2&&R.nRbEƬL`%Xy.˜; T/TN@L0V>B 㟹`5jP,uHAq$ m .JbPDH2JQؕ\U~Y -̇}]Ӛ,9pI^5'bcD$K"D GXN+c1#9 $a\< @U ̱qrrYkU^FC3L!6NFcyDH!,\;vz.%<0P!zGp*d֡oJD!;#`HtV cTP!E B1&@E4XYdD)9 > uSɥHAy']@Kh-iL4D&|•5$Pu cQW LJ'vSDbIJ7ӯ+GiuEx\n@ӇruEAB.t!4&ݠ4o/a X!qۻKn 'Yt׌:O_ܾ#%d^^LlǴ~pL0%[Wd +`z{/~U-;q?\k%~y?d ޭ7eYqI֝R9]q٭})2ATqoƺTg6u(eTtBSK ԁ[UkaH][BEPtDžj3 iADũʳ XS?|W>\-rxu}f8^RN UCh˺$Z_4bϫ3 !<-2(C%Dz?îjEAHljwA2}-s'Og2^N]'I菤%-tӧנC3&K :*j{p8 !tUbG%g>#gi/zjqY.2/*{<#Gpğ}CKhy݃0gϳ]:.|rBWN,H tBj24W~ m =ZpRB$_es"xRٜ{TT7RB ״cOc|4z1cmhwû#O[E[wsCmBNbo}H^ .@YeH8J~Ylā_JN5xAWR=jkb 9O&K%|ㅐJ60Y0'Em"/xj ʮn36S0kS3]P#p2i`+,vIQ*\#I %׻;ԃ@qJh/#Yh&tpJ:_!&RTM{*O__$kĒ%o/6 ޳SKsFlh CIQ:)@b6b5zOq8,ez-?<\%W+re.4 op6(K0?FKN6R Xo7zƮ fy-m7>N&Уe yfͤ9i=H!)O%y @HnqJ1i #/MZdLZc:O 3kŘ ! >ZxpFT P6` &' s:n+VjA0"٠lTxےC$ԨD h-!!z'2E6>,/˹?n D*뗛G ?LQ .17}jrIr57.ūM .Zך# ;fT*Ew\ۄgQP--_Y|U?΁5YQ|Rl)j6YqMPj9xcrK@џ,tWcjSy9j7n|ae"(EF(A=(b-ʟr] Aj]d.u/ VBw| KM U}ܪ>EŹ5~bTr?N*.õ#;-'g##oqUZ=w(Qxxb%˳w[»A,!gN`#RnaoMDl\g^pp g%0LkpFsv:}}הuǪSp8jD{[ Mkzҁ 2IGʮ-#JR(u>QK@Ф9Jg$E7kWmX:2+fXg;A j\R:`'|\<8ݶFք2W<8={wUJRrԍZr 0z[R*pS ?L ùԊG(ry&SqMт %+cUR#&ϊ:pJX_E^/JzS|t`q~{_?/갆88 7g71xR#Peeg<}SD)6 }ӺJ[S rNgԱnC:iZJ՛uki!|?4ջ2:/h/! w5/ /?ObL1lƝ-f Y*I.~q}Z]xnuZvz=q׫G_o |zP)ɼL(Li5(k.y@t( ՗ت'溿]; PٵW/6-;iõMKjO2\!p*sZV]l^jMy6v-jWdacwh{=cM@+ワ-Յ<<=z^gc0I}q_kEr%c`"iSR8O>;}J-,nN^HVJv6N/p6ҋ'O!NĉB$SfRؠpӫ0j0` O$u(S0WeiPq&(iR Jԃ J rRfJ?MNtz"o^Dʐ:xvؖ68ܹ}yU #iu~zUʰgB`'hŎu*%E5%a@[N>2u j b08H]^h Ru:9~]4 ){O7wH}H"c]f>IH6Iy=)y#ՋnL}ȧE|*͉ E(WiѬSUP!Xi3svEpW\% `umFM{W#F.GiFB RK#Wzvcw@jSݣkfV<^է[ I(fX_t#V˞&I 25@p<Ǖǁ>:0g8)+5:5 Aip1p.2"i#9 ;._e'#:ˢ7/3H9|9!DڌBrG=GjEk0k!V>Mr%0;+%,jWb,/١^|K&>?lbˍU$7L2XZ6$~se\z8WU %dLeӼ-s"rV(Z,//R`#*r-JX"_;O+燋ncpiݗ ~nsZ5k[yRx o8KZs!WI`xټ(J)J"K3\(,i>="U>,_ϖ@=NU/6sB^W+J3JaP$Vg9`P& d`ё>z K٪P\GTpéѳG+ɒ”(=\訖v}cPW6C^R<9R Qİ<79e T(`40,J!-AզJO-ƤvPeI.n.1\p# .12gt3YRF8AJHSn "$)FYM!Мj?8*n> 8C}"Q$Z9+vL߯^^崝 Z7^//o_DR1TX^ξ~ez;g@De?!ހS6_,7·o8Y{s.L'k߹fYok VG M8ּ)%HUeLMj2jcV&>bnDa'Gmg K%yMw&Lz}EPr9 .i\E&JDj_}* _k7.UŌFqIaF q` ۛ0? 2pk_ jm|wW*W=A] c[l)?JA~t}zՐX\/6ݷi}zXf41RFt_ ./>L?n__IP*JcS4Xtd֋[T2*.i|l\a7{M/yG>V+nX "nzh}3.o~u or**AORaPMà-)QxؚBL NB!'9E#N.BU0tO0bt^Pw3;_$8Xh;<0}Lf0Q6*]-M=:ju+I-D@k5e -0X .l-l)1J222 el 6F>l4 vѭ`E5ޣ}- E ^ (;:|},'Zj5Z+UR[YO  %&TDrXCսU']ĥ XPS0O^" *eצ"~M*ṉO-#zRyP'5VkZN<ן$0z_R݄aKp00vIg qkE;$=ÚJ\U߿?i.nů dY\I|AsDp &Og6_AW=OdIGV>}SoܦI )Kfi^W΃V[s%K{Ż?Њbez VCHeGq՞UFYt@Be#t5:j"$$]-[E {]gP ]ư]ohB g(5e3ш)_wR+PHt=}ۋTw߫S-Qv"R,QG3Gc2K+ʼGu~A+ӗM. z}2a Ѭ0qBEK. BY)-Kp"dG䑑mClvNHoT_2iS?>O#r~w{qٰ~$?o͌l똲mkȨw 8zxJ ?(| "GEiy cbn1K6g'K'#>˩a6f>)Jr.Me#ê b+ "eL%6mnkPqppq!ss/l4OJwVYV\5Rdtىo &aQ" <_>tb mf!<vHL7.'Mr)P;1$< ~k]e|Ϧ_ܗ ''!ѾڴDdgǰ3hKy`p;Z菛xP*]!g5 T`ʥ|G,ŵG2z#8Nȩ,mՀ `9ABYD (RnhfIm.Nѡf: jupGِ)i=rRwDxVGڞM8Ԍg8gHp)q}X&tyˇGE$%8Ky3{NT4|cx=Ž w90HpEݢA-c;;%7 4/3T2w783̑ hDRj"i#1MJ[!3A,ڝwS=SӬ%apsAɸr7I1~`MEA$COD4.J,2}poֿZY9*=zĉo U-,1$ r/2_r `|g 5 Pj gUy-N(P`+B٦:ս}bS0 TԌ`8b|B=9şH9;}kc`'U{%n~[f\z}8Mb{HHepܛOQ2evX-;3q)ۙ>mBJ&" Ba~m W y0P"e %JUKRc2SߠZ~ i~Ġw/s{b̀1huĆ N,gyVLYh@˃dpZa1еF_'6v߁O1۹9V3 ם.S>۝8z`6aڑg:ނ9ʆ ,>B3DxG%H%J )y {IIR<rx,tWUuhut,ګCYLYekO6m&w"si Y =9ANt:\'4>i ‚8o NYd@IjW]CFUl7G >¹VM /*P:t12B׸0Ѫh;5i7 nrrW@!dY~xG 7GR|PI}TE]AWp ˜*$, ςJV; ]`ۼ_^łI9]R ;7nȮSbN\y|t(Q>7p /YH޵KdjOYp49<脸]HqdIω Zpy8h sZ6,pcINDtK캳qMpՅ9G#ۜ*LR+\]_؁E}1jo&GqYDBiOxH5okjN;FU8Jcs;3Bni\FK_2sX#e*HREkϳA 52sḘC}E13/M>z;jB-֖hVX݈r[M{OA|@kC>N)wkgMOusD-t|NG;MGiNg*R&M_l \6$tn YҚ 3w 8q(ħu\0 XۭN>9]9S+HMc ;ki;`ַʈ࢝˒j_ AT6r=%Sl=7ϮaҹNkL?2}?k2`I)YZ U諔*(aZ-T - {Э2(!NYstK;w$^Ʒ7~dC`$]GřW ښ4c:9`A-z<awY 4(ِ ʴ}jkP@ %2o+sĿv*fBuN`sB D2InlN“߸)+ى`t>(_M&{dЖh 4-qZ,c-?ՃŽ.dG3\D_m;ElwbJͷ|6D彝#w7 x h wDюMnLQkq%[4V}彝\UZP18՝Wz|!#Z;Kˬa9+ջμ¼YIBfȟR!eR_pvVumF(NOjR]}rOj}~zv|/*m `;vdK-nzq#6q=Sw>"<)ڮ =] .ۮPl(iZCBx?^UNbjW2M -#`@*fV3ޟs4ؚXFyMߜ n1BaXC'>xw6 ?ߠ]J6X!`] ̺>NW ?x2EirNPK , k^q$t A]g[ ?NiQ7ߛ̚2avzP!O͗ T 0N fʦɣ3O7z@6Za5۝: vƸfKZLe?}\? :|ùRzWW%^ ȧ *}pk,NݍGc\P QH-AviF>LsXo`}4҈!_Spٻ3busPQ3V˂+CVٟ~zBm : {O>1'΢ᚤ5f%'eɼ()+h qHLǨ= zp:U%r2:9DGAJe]0Q1c6Ъ5!:<\VPCšCnᖝE-HD̳(0S-~%45Cy.PCY<"P2cPAP OG#^eԨP[RĢI^0_Wug,0e) Q)tNҠ=5ѳo/H!VyP:5 ֐JU:!X w!' Ӝ2If' _!jҐ*tYC5 >%aR"Xp$)Ĉ` u^L1Io4 UB&_£5icts%+GE) #R;z,xRRkQx $Uz%v?R 8 #r3 Q:#( %PTDLlOA 6ʚghN. قqeR+mhh3} $2xYBg(W囐t  oZ,dDm&ZLN6>zk/*<+mJ A)e ̤e֘g|u\o *XW;eqa61rgl_4Ģ-_6fm^eq[nŒ%k)8\4gq 0䓕n7v[u\;Wug>m7pCw`^Yӈ&o:'[U{s`Ѭ#;b9jPs0dܼ Fц ńWՠitdhE|'ocC.pθXmPian7sRO%k<#=9 A/2@M7[; ,5`{iѝyAy&ȣ, sASUsb~-Sk؁f=-ᇻ/^ a~ʮŶXS`0O22{?<\ߞYٟ{ӷ7d (ms_ʹǶùW1fg([ :Y<ωEJ|-Rk_jkN)OTW< a'q?|%\a4SO.ԓB+&5Vogw/9\mwUʛ!ӑ~V<7I1_c|A5FdMI8MFx7 $7i6>tVz`vn%6㊭zl±X]pHE Kmol{Zw5 6YR"?k=WʼHUE2/U9!uu4X: ̢dzWPQ2 :PB u9q)KKc090P WF>U-y<^=P{WFU2`i R4e8\J" $$Rz3`'e Wj%RWL]WRs aA(ցOs-5J/mVDiF2(ܰ:?oC#^XK2g!FQ:PMJ6AyPa. ҖNy_)!qD=:+MDGCz*YY䨋Յ$5T0BaGs&2 .)0%HCg7(\Nw; orc35Jh 'n..$% WALIC *LLici7h $QAkM-mY*M$r!x.-nVy@Z%Ѡ [s<_7l'qYCjxPn9]5T#jYL?_>6JɤҜjE{|֌FzgyAAhٛ}r#D(@RHh " ,MFfc*ѫ* £\ =K"Rw.~w֛,]04J{o٠ѪMZVhQVl`ZVJ)RWtV㢗fo(An1ަ030g^o/)WHCIiʌ6MK_C_(jwRda;#_*yYMxI&Y>i>Z8ü$(b6dVHQSq`LY7;rK0p֐y$CNsTuG˩3(z&Ԩzp&ѣyvh7}wE2\j/Oϕ6̨u F3 hUFO>zxI#h ȕ׾ |@4m7<||qw!䘜j{WhT.gBc<`jw~q⢶F~'_='u-ٻ޶W’~1>OZ6E/KK,5,)Kԝ/b##Q3sΌs°^֑[2ZrXa$x0=g8r(%2m ( Sapp1Qa4{Mܼ36g^Ymz8W y9Z4wl\xgboNup/ݘq}[t$5Fz!01(xe b|;o "Cƅn?~(j#wkHn)=ix&S~r1a%#=IԄmz#%DF`y6#G?pU~ޟjK ѱ"tJfzd7_^ÉD~i:ޚI8)k y]kѥ\5˘ [ 3oהhtMTʌ% E8ٕ5S4f+ݱiRqg3AciΑz6MōI;cXBraȆg+{^n{6ICW70Yɒ`UЫp圦;=SX. F1%/DzZ:ݢ@'B#DE% ]D\ٚjQ;7lhUc21)E\M?- Hf=hG\aR%^[K)w /pwKVχ;1h[97ݒxJ»궈(7yƌ lyPXF&oMby?Hk&Hx} -vJ-}f?)`+j?%BT#7rS-|?$ %2ԙ5@$蟽S.DmZ{89=J3,⟷urPh\xik]8NqЄ <%ռᬠzA}AYW3ZNM9[UוtYNrseo%}'H@ښ!BhIIt!Sϖ2vw `J)pA@HHb G3eTe"%Ң%Oʓ-#T0l֖ )K*22lWƶ .dl6Ws ld\8Oum7ۋf\ hUFļ1Io `2Nuz7l`"XP }) 胉1vM85R2\QLj0R)Í&^QPz8rjCaiHP #!R8WRTyq߸Lx\ Z–˲ig3o!L-48fy۷7h>їMGX'4]a*gwYÆ>mz&ͼonxƣg99y y_I%9aJ~@ ]vl)p`4硎SˠndҗC`T:gل^4>}ayX*9h$]ղB^D4*0+,pݑqmG׬ɈG8od*,\ε4 N/A*8]Q,\XB8>*ŔՏ#Zu*!HTދ5T:V[*lx2H"yeRu IBux6eY?E5Ha]hlRT '3qڕKxzzSVpkıK^&b4.[S3r4Ӵ&Nyң4r\1'}63rI"gVZb,C?`}u3dNَcڰw3DZ qI{]=}t۞ jO (: buFϦט0oC7NE/ f ^U>L߫ׯ?7קQ 7.tN>n??h|{W|<7+}s Oo_nhl u;;&fztxGsbE !9͔IsfĜFmtYی#;fZ֖z3qVйZ4鯗߆7v3{R uv↟mh ?mXkGd{i8b T;Fe(dЕ9} m֡(/Qtuh~u70޽׋5]=h2&7_oF,*T Ռ;[g?,]xpu]X4?$= />QzFsv5~njhţGɳa~߀n>1b#Ho^ouˑG] cq'n|%Ɂ$B;|8<a?7c`YF&|Zh?@PE"'xM$@,aZ;鼍#2#`d#ߏ=$dnz$"h.Xu4]OcY4q?_ GaMt=UjHB'יg w(Bb3cCQK&}%fhFHGhX ,q1HYOIĜ̶eL("F 0ޟX:75sBCȲ *gƄk\xWxRK!e%-R'-Ev2b9³ʈם4~Lp0ޕ&Sm1-bքS^#h$I9p#Gsk%$:*4zTpx^d-uzpťIΖ:zr(rAOe:b!ERA eyT(9 ^ `K& {G`Y}[F㙒d=@\<#K=bʟTRA+]X < tbPcCk<" .3ɑǺVs;:7yF<1zNiy-LU)i(-.#v}nb&m*P^H[ES"WS,`FcѪx-/1sH8w\P*m(wUǡlqkزGyve/ˡ ?>:s6<{`U])uΚZ&J9?z&]dQVƚ̦GÒC$K$U:w01tz*gFPQRUʛcG lޞXD~|Sκ?4-p:\>r.n:B2wpup~hdzo{g:pr;~ .qOe{]T=EUrOQ |չG/O58٬P76Œs MѸ 6t"|hڑK ̆K#X6;XGe;6RCЕkhKy{ ,Z&g?ؿQ߇3uמNqzNkxo +{q}co\_LiwKso޵6m+чL縵'vNt<$ƪeI%'']D˒EI.j: /g]]uu\1A۞o;-?VǏh>+) \Ȇa~JWcwn`E;K ?Wf?1@&<~l3`~{:1ğ ?~ϯt/֤qvl]t%i>''ݯN'a(4OnXx6U 'V먂X}u&l8Oz|ʹ'~{һϾ\Fb^_A__]KWUe@L).ĩ'XOrݱj3BT?O`b_d咆d8Y*f;fw~+Gj 6KyG<Bi VV vKC\ H -Vـ "NE:ܬ[ohSq}n(e߁Y-z=2H+EI:1Xy@RC@BpU~l8(ن]Bd1l: %b7[ B3e0at[` 񪽵S@Z6ֿx]oͯ #kܚ_28Y!jvHO^EX )+橐 UBwOq\?qq=^;̨spF6UZ̒AU5B} _ >7W U)_ cMIbպiv?}@s}m`D\`Dr2̕΂d!8/ͼfߢXIs'#ˁa0 QKaҷ)]2%N(åCE Tճ_ zih71+>Y\jz>*|F^B1("<>:"?N?V(F s4-uwDLЋP@HLp]0E~cj+2T(NuwAUHpgԱrL]均yqP4MlBNF|}.ZѴthF*\єM `,Yi'BFh,%OĢ0h/LC`4V.3m׬HK̶ A^l/hX7R'~74WrfW&}N8Qk}dĀrR-] ko<8~mfu˙N+v ,#-r!p}0 + SJQ=46E'':Zmitb!畘)eΚ8TR.;pRv2;6vnRH2c6_)O榳]|]/~̙]D`;eZe秚(KcC7cC r][*T\u=_zFZq,[eoRT䬒pнoYTbmЭuz5&+X3^B2HqGCŽHcEHXb9Dzs\gKv+"_8^0 Tr TIx4b@] [7f5^lبxDQo?cBQ2S~}x0r#K')f@aҏ 0"؈UXGsG% [A/2@V~d$cu Œ萇V`ak \^_(Oq$nemw)Ynk:VnJh{-@4(']#Q>=;?>D̈́TX=<~1zR <9=]F@@"4Mx]2ad~{[>ؑ=$Ҙ.f~ zW`#'Frr{8]+ *_.Gz~|n9Z &0X)PP\&\&&21B+±& wzyT(gSP4zKܒM"_Uw˓Ndsb*JyLA|}y)+b"E,9 4XN JffܪxY*PTj$/T y0yfgЈo.ULu -fFk5Ab*H iJ Szzݒ!zP0Ʉ(iQghsi%T_؎Lʭt/O6WY֨J3]LNhc ;hC>J jQs_NzcJGuwy+!iOq:5J:GNxg{a*cX*I^ʳ z3#ߤbꮸE,8+6 Evl)&Y,yKL֌g){}:XJ%|[% JTj62ĝ銧{q f)f}U*Ic[/ Z?_l]L.K`zz^7t^Pǵ.pvYii|wlH(uL3jxCp4R߉8I65AVb"`4m.&ҝԾXmjg;FokFKQ=H(Hj&9 σޗ֎q/ $ZMl7)"qݻeHy[w6{?5=x~p~SO\A)qYzWl$ߘ_Íه߳߯^|Gz_ǍGٳ{t޼oΛ#[~޵w/lðhnKd0EHs~vh6[2dDBRI4_eyEU?#k+dԆoJlt/:Aeq|4i0':>~Q|?8m ~+|1lw׶wl'5@5psIS N>6~~6f?&<~l5M^gE՗8hğSNc5=Icܷ::i:_G?aP33Ї^g ӓߠvy*! ](Nnw}?ߊmz}vVv}{v6QvWW~6Ѯu8a?z/r0桏 Y!EZYs+}s{3~!,z&SyJ DbIQV *LMQ e}3҈#I5G9wn2s4cL/>hF|?(R~dqP|^!EJ$պ`OևX&zE.Nպ՘N?\$}j-=^}H.'>c#>cGP,6:# "h%<\jDxBT CcMІsh?7 H0>a}d$FzXgt$+@vCmb{qb8 p!`\+ĵ6 2D&qlB`d0:(a+44rÄ |KX!uooLVMR^gchc43~zZDqKtd*L̤Гnhl3w϶GGCpCIn~7@Z?֞H ]TAĦ :x&1{=ZM޺!k,{ ͨLP 6YOS K80d9 aAD.wfL3ʸu{c5'!Q Qk :x aU.l(t/O6BW&.] db?V9ASIe0^^dMٽazA'x4<!jֿJr)0"#h)*CFQƅ"a$&d2"|hn`aN!S!ƫSހDDQpb#L1BicxbH B`Pdy"u^ڈ #?ΑЙQD4Ҙ.; ]?ek9MI,Ŵ..muBq,H[I\Iyńl}Bw DZbzn,,ebFdLGckW^(9̙l`CʊR= qk~ 3C4i _0,8LϬF(DseD $V(q5ZɌPz}|~DE3>pv0E8r|@ "tC(𕰂@Hb,˄9]NʊGb(UgAҐ d <X%u w0 *k_qx}n;p7 q.ThVf)RLX"yWC9УgtNyBP?{m@/KU\O1N ^BBIOqFLĞn^K=ŪŢ]|}Wc-3%c ,~+_lQ.nnubgX`ӧΜBsXgK{{,$)֖H~A+XxsGr7-~uSjގAFlD64=7%eaKSpK&'ZLc Ӂ. =Pz$z腇m5E:bakwyG?\re{=ޠ@Jq-bz>)g\'o[f8Jc†pd}ph|td . amR׭"2X%yid2ÒbhX 46!+_[o|}1]=׊<;`gȇa fZgkNS\}y.`#681}礤JYCm3}0I`v~02§[6R/ qQK)-̾e(MA /mNb"|,VDK;Nhi+TJe-ңl2Ҝ,r_$L1.P薩? mE7~`qΆ VF[\|y ??pVopW#9l+\ 9u_A @Muo'\ux^^*~>' hk#pTÿ́E&pxurhjqq.H ^ESO1Bt.hbH9Hk}uw?}GN^A/z1,}Vݟ6nv$Fy~~E%D=wD@YpHX\Tpuol $@%Bi-Ad5-c[F 4| bֺxYU[?GI9"UG T * D,謠E^PJPcГTxw)\JH|օ%(9&wꀦ+L؋h,-Def'2^$-Sg( !S '7 >6O' -`WPS@kL`m𞘀+B[Ng};`;L0n/`d5^R7'[cQX D+Kk!3 &2JcJ4snrȵ\tT1Bd~+Ѳ)jmXz'Y4, ̲6y=FayE:ZK^JbmZ2>ꔥ7RԖx9lB-o(SYqrnm4-@}R=!(;wPAx afwۑV % 0qP]XɰJ-34`Qd9'9`Lh8N$Gx<1(AIiMyXMc[geekOAQ>)H wAeb9A=Mrg'e>eјau.؂lDts@ W9qn934 9How9NHjo.)e ;8-:ԁV+)ptYA_vR)m떚{C`ⓐnH`XaLhmD-] lg|-c߱׃bǘN;VnI)eAhuk&TO1˶6S;E5Ŏ1 [+ëW #Nn,V|pTrQ 0EM%i"dV؍!Yק(m:{4Vl&#]ߒ20iY#T&%Q ,̢ȏ:& _.E+H2ݖkN pbCi|c.cǠHgB]($g{9hU˵u<<{"b<5l<4n^k"Wվ9Blh //" N̏LZԦQZ5AP˻uaK^(=Be_3Jv'*;QYw&9/ҧ(9)7̘°Ա0lJSyM/!1QLەqԄ&hf8]k8M>rsI^XH#GiRhkBhjj#zjRslK^(m&HW= ۹,cXJv Aҫ`FhDxrg߂CHARHFC(̒B(b1:_ODY} iMo ou9D{  tv]ޗ3[v˺;/ݖu^m/0Z}g>\ɍb1 [nBl饅pxKH=>im'Cz~&{!ë䏟L7mFc=1[ eܲ:F3ђX>ۖOn`Xl7ڰNmcc}hޞ8ju vțt;Y˭No9twP8QcN[L hNjT%^|ڋZF-q,z/+v>j567|v 里ES91w4ib(ƍƹ#V8xTgP7?]u߃:>:f.߬ϟr\O*dz|qf_/xuz󇋫z^wp_~'GLOZN==Ez l+븟&[f((}󌮟}Iv3j7[tO4{4I= ͺO7Vt?MopsLkn+Hf_̽*R4s3sϖ|4[CRm>:VZ#L3orwݩu 1,ͦqSX HMBNy-=~eKyaЇLˇ63mI`=g۷,6[6a 9ruȏ3F㴰tQ !/o1mtZ|mZYކ#=1RՔ _7ƘWٔǡLYGU5y}NMj?^QXLX8嫣ѯ>O%(ïYkY&kδuxS3vZ17[!跟jR4TL + j!ؑFqc҇$j S?'$6=ܺc NR >8y57qbpp|3Ot(Ul2gk ^XaCI^?*]p%(h\)%dfԏGߧ ჵd9[,.~st4rdO=sl@=v}#D ٴH(+Qk4I(O;^䂎 Lyzzqqkiۍx9#9^NY0x1H4fgƣ5f=} V'yv!\ߢey .DRҭ,l8Dx -J/tb=zOQEԶdg.BbKJ8A6o.0\\ٱJHeĪ^&6-a> 'zvP~aFBBZU\O]]O @N<`n<~w*}mz#6BE޵l3ҬZ[&` B8g].>T}A&UaB IaC/(ď>s (r8:^kMhP$L۩^K CurwB#sMN]h i=deH>&D6xFP|ۯތI5FM|4~woģhqׁҙtAz-q;B;5(!5:|JZ&bq5*\3)hEt?ƪTRl/J_*zAAՎوXF@iYਸ਼R'bzP+ٟȫ :W[pO!Ib,vlȝJA*-(p^ɣrE^|QNi`D7/"- 33U]׵qpzPRH믎?Jf-ղBH"&{`zz =עfH<߇mYȥ%SϜhyf8\$iѹ⮲B myǗ#=0O -zXMC5s5l4j xR.Ow+ І-_߼gcTz{Kzg~\Ř*"BU7A?_~~;%"}(p7Nn.78S|*\<y&z5^n~woƌIq+P#ƑXH?z3>1M$f>ތ2&(啱.QkDXSI'R ׆(=u )ͺaI2"&G﨏)HM5ExG- r1&$#x-!`Y@x )ƒaP9ӡ5=ĢENŚ'Q:~*AFD62D˂D3rRڹ ^"^D+1-%J3@ rG#`C) Kn'v:ݝ/9h(qba!RGsAY>)Ax`򂃂&\D`TuA 4ZO /Ak?k+@Ɓ $dsF 2].%q w)Xd U?/soⰚbһF K 5%N*Yk<  '$.8VHL$ <܁ 'S錖Běp< }b.njˀOGsn4V1FTDW#IfJ#"2bacg2gKv[v{7((UHʊ"2/} vL:1HZ'OX`g=.,mP v.S1'/|ar`,n@= :8 BKOQ FE"J:7HP1`+4Vj%Sj߄h+4jv5&fH,*Pڞ,eBS[.Kdn޲lL8>xM,AuyF$É؍GJÕ$Z RK yǞAYIÞ'D Ƕ!s{n$@+-ٕԾD9*er򁔺T4RmQ"ȩeAywۖ":ZЌUb@N2,E2;@4┨8-[౭c$&^ILޒ:w8s .+&p̾%>B7s>pn {!g_2DOk TwIhftW`4;M6ҝ" RC{!n{{ϴp6VkVk͊T.`|,`S@VךZ3ך[{E$M0Jm%>TNfؚυ/⢈I MQiBe5;{ՠRkr8lM&[*h̩mWbb:dM58 m3L ,MbȂt4QۺH=O =y yEEGIgJ#%rvK#;rlH 9U q.`I3>kdIU#%DJ NC%hrU ai;5)~=e˙b-g< m7rG ֪fRsH bkV}J=(|;$Jڪ6W*;<CFF]޾A׭'ܪ6ag1uw:k;Rhnv蔖{N|־cvR]q0ctvCR,YJXoPdI@`vl{zߩ푍gך [@4p$ʠOWz-PKIX [;U> f s鳒f6/ѓAt7D 3y zp*7.J/z( @SҎ ~zЛЃ'uRܨ͑X`#b/=%Z$ِu5%1/g3qǷքuC/~:;9%#4o+Nћ\MW!r+mWKL;kϵ4?KH$_A\NI2K2-3&aI"_wuoS1HOðllF]!OmM ]c;Y ϤK r_䞗 N^k,xr֨ac yՔcCcGL6O:t́:#dzd) V5mop `3Y1O\'XO mwYŀ J&A-Xlγ]P|#89*52P@n meFeylOgY 'R~,ı!P=f5{x6lVB1J~Q"򶋳bߚ)U[Q[g3u: +u4svUo%4HStJBCPu Sm,Y-B!RH%0e嶐-5D.. 77~$1P .\A#z(,XoUic8[Q횧)+'}\gn\-\#Ӆ7ir^w#Ԛ Y<-uWvMRܑd段K1~`|ew_ƭ\ MjfĽqr-qg $9K' ˥?H'lNyyr$Q x2/{rKbE:?)|bUo5~k[y=}:vgI'%ҜMX:;Ȃ'̧O}:k=l$%20p:2`.xCeC!9!M6x'yLLk):>F@$8ck'dD}اqxU0& %GX5 J{6}U-ʲM+RfhaiAo[NDXChvʫJJbZədt&J1[ Fo.\ߐ8h~0v'ůKz`vI=mM ɫʏoYўǓ\]q}&ܜ*;`퇋 *pܖzZ6uܟ]qyT/Nj/ΎOܾb*@w|cs SZjFH筬mLJ)!K{TTUf^"g lX2 t޴sB#J|p?߾'ӳi8|<5a켭*[FL10-kM  5FqH0 " F|J{Sl,톁0d7Ɛq22zʉSlӃi׵> |+yWڑȪB آ\RtJeYQlҪ%foxJnDs F5 chF'T1Trq )aDKYq%*Ojx%nٻ߶ql0wga%$ŧ~(NiдhB[rfPMYdHm<<<<-i }ڄYNg4ޔLdŸb>JqlL=N* rI =GF{Эu.jR5S4`1+> cbݡIi!(OFb r4#%_)]gZ7ZUK˰) =3@G궧sanN?ԜzizeYxa2Ӥ̛LoK'?VV$VDaPKGK&bJu(`F$AXf, ܂v(PB%H;mH=!hӑID;Aa:#P)?wj<IŤRLaKW5i_,=~ëM S\rEB _: "FI %(M"hl4;2 HD\z*CpL*]s= FaQ|(Ι@ JwÍ(yBs [$b, DCה/fr(H@s<@YoSwk:_SpSYGTL& !,.Ue |Hcڲ'Ĝ)˷Vlܯ'A|IX-"@X4Uvm-WAmVYUҲ"`? f9FHJ9+:01wܩ@Vi `"Y)HmMT8%,Pc~>2Q6RHOc$ #Q f(ցCD$AOM.={'ffQa8Vt~xɼbuJվfe*'w+{KʋujYaQUTZ(F[!$H1 3 cR&&D0, qm(B4蛮d(+qd,3K5 dflt0dI!|lc4iώaSzyv gre5䂒 ?8ػ颾R7)H0 zی"~cX I]<՘J4ZI}~So*l[ R Fc $?,Ay>&Yۛi<_|59+ceaƿ k?;4q 1zϠ-+0OQ哦W[;gf2vqtJnu9+79rD`&.4[B6qh}۶Y:\PCx+X_DU4S$J&2t8/(u|Ӊ9'"E%(spSEL_9EEʨ4khUcAp:Ć-~((8#6,PY-|cjcDD"J8!VA@Z'%i !1(2Y*YAh\yCi~fBcIҕʽ9Rܱe?H?"MzOF^1'f -9@82LLg'/ak|F;svޥx챃`Ktz陵?D0rYgO{MH50;k05R&~dZ~"n:O0ŀ$wyNqײ? |Uݏ7<1o|rv?z}ջh Md >/H1;.,d4E~ˬn'rx owqbvMMoЛ>n!]#م{|%ʊ4Y/ hxKME{ZGi<3;W7a֤p)o2٦;^ $Ƌ&KtLܙ>O2q 5`66-.QmrI ݜCS)Mtd 8gSO9+ j'kR\YMWUlaG;@ggE= s=n˽9W>F&Hv@sH A<|D HQ n=~+,,+åNWTpu1I~uw~.df zpO?&'bh!Î{Gt{᎙c^oϧ |6Jߋ߃seso)ie`>t_kX?w.`t?>[XJۻ:z7vlɻf<(}>̅I&}_e|=ٳ^LYs@Q6out=-R8v& Rd0\Madw-(yhU3yoyyVO jRz˘;gtfO9^48З7ab"GU@c8a R)v8UalrAUicij)L:ύfv4bI[oGWEZ4t Vas6 P5a V Sh# :A0km±$}[ jN\J>*hF;h [ijr6Fwqk{X-h4f Kݐo~/jF?~`[m]6X%2Y2cj/-wz ўk[*𛎆95&qcW fcx[L]/.Z`ϊ)x@<'[+ز;?6wwRS: h%:$ #>μbfx+ۊow0 vsΝl .y{ykq3u%(Qвrp35Ggf2Q(8z۞Yk:f2sb1o1& ODHzTs)_jO(VP1MNc U,i.ٶ"mYm\4XV>k}}m-֘YYlV[fy0=\+r8 ZMbFo6ˑ=ӣ4ʁC$:ơms4xh)}8=mRvtnh늫Fqا"3\6(sojevz6gρƧxmICavMZ㼵s^n&ý!IP靥;mz[c  fF1~S|xUt\dKhM_\ G:}q=UejgE>F Φb6]<WӠoLT!Z`9*HAj,|/Z2᫛gXHrEB. Z˂Z&#*h1Uҏ+7,U~[|v^IG9h<=?]X j)YD%È5ILZ+(>&7: rO J;ǶyZTdvhk['7N"eΈ̝Sy3i޵FWOU k^%(S̖Du5l AHf^2RtQ51nWfqcxt4BX3ҫo5R8y$ÖL`Z" -h3sl#Ɠ^UpiHuDDX{(tmIbOFW.=/m F7 ㅥ?yE? 徜uO#o`z=tꍃW;Wؠr)!@r!<p2}b[b;S<ʔ ?&^Ddg O1@8@Rۨ9 8YpyfCuVQf2)l/ޜKLmII>;69оBˍK?%(V#ON)dNsđ\01/J3yIDS*X_+L?,W F_sj*fɨwGB2\e q/o+Xe+F qp'@%ODKnW:Mk./hqK)&A$ cR`B}&DP$$,BR :c7'8T1`gO#xAVE!E>+|2~a@$4JH@(!,quқDA 0Wb|JP)#N0[1+mB{!,*&B3(`EZ%N' y@E%>P0f~DBc@Y4f!-:XRHF 6R(H]g ÉBo7q#0KFKͨhH)xla)l]o#W ˓Mx3 ȱl>e11 lNՇb,S_Ud2 nT%Ԏ DETq_@׉B754C;[2wmG0ڄ>իWbv29~QuW ށV Ip4U ޘ?ٕoGj\&G]' c,x'X \Zog=Fvr~'6&hI,9 (4d<)q lJ8Ǵ:59ъTьh^įOs8o9\ 9gO.|(}Aɉa?f%$cqtdO#AHRjcIg98ڡXŮ4mJrM'NT{2srdsr Jʖ3 0yr>}r }pVz=/D/?($6ܽ7z0nh1 :(tzmNzo3=*?-XWxqMj߻WdFz͹⌂'!4*'][䖛_!Hdw@|xqS-dNg 8UNn.o/* MތX0\ƑqFt<06^S<\O/N3az!NQ2MQUgJߢ~Cf7'S4}m|w>C&E~0׹'1qZ6q#ݽpİ~9r*7/P_?=L&3=Rpg1o8/GħӚL8hIrG<Ĝ=$Tam|3k;ޥb$ŵIDϏH$ H8i6I4?hŽMTX{HvDf.P X,\IK\hi~ ΠSC*_n6*4`(|4nT+=0WHEƍm.Mh%Aą.(Z<()Q)f4|UM0S +-jyU^/~ۂmH:|.7q^IVBphTkQMef 'C/qԳ'BFK7챃1&y^TC;}efkJnNYIh;.i Q%Z47' ׼;=`>B'^nH*;C^B(ĺa3 qpHTu~ztL:rjZvµH5-+t} OS,H0N2=uN2c5N|as陋R( +Y-IF2l:y] .+ݚbP":MQGqEVv[㉖j6$䍋2O^j~QC0Z1Bń&GOkgA4 "̥|GnNM#*;P~N_h-pgЖ77fOn.~K}0\qDvP%-e+[NwQ+dhPne (:!h$0 |KCY.eH@#_$1,2.94 JE e1q lE%BK L'PTAy䞚ܪ#ceCydFfhP"[A邻oG nmI=-ꂻ^eƫE`%t,WF3J^ɀ&g`]bg'#.I"53gG$K5HʁqM'eW.[fDZ}xcoej WB^ 4ڕ6k.78_nt CxGQ˷AHeswC5:\'R+f3B|(q0i9 Vʕe:ěyܢzBN>󗈛druy{w7Yh5!QwیEO0xPYݷwӫˇy#$vڌ;( maW%`lkgوC6HGZ?MI.R7^ T[dlC]A>Z-y5, pTOTDӣd}#_%9VTtv c> d9f~L?g_f^w.W55|*uA7TV߁svmC** jqj6hq$䍋h#N.ɵ!D-H`J)kJ$;(#Rk2P!!o\D)r gؚ8J߳+ί`or6j5s#ieJHؗ#ZUD9q}lw?8ڥ} 6ş1D!%r^nJo.G? T-GOf"wci -3h3zqNxe:E4uIjKFFY")(Q 'H|]PH D!Zٻ7n$6-;`'$nMk$ߗl=GOj d$_dZa"FKYH)iL)6 z#@zBiZ6Z 2 X˰KZ_n&i;~3XU<և@%T0D$ )H6!)Ύl1fT-XYP\L0ⵢ\Z8!Iz%S[U 7Э'+'֋aEzơOdTU$ 0JRۖ<ǒmKA@̱ ]c΁! K!)J6XQTj$*]#lgp5(RcHGx Ղ3>ys^sԝ_U" U4u=81JA@H!@2-}ud6ʹRXpVV1K&nf$P B&gaRn&{n*Jh*0@iDܢ(Xb*LCdn⧊:d!>t)wK#vҐJc;M BXRɉ+A83m؜Hdʽhp!*.xMJ >Dv>֚5~X{ JnBxhkЧ&A).ZwVdc򉗨'sd yr2!BN.?t{e߀]w:Lgc ={)taV=v#o}[G9lpėV`Uj{q@F-FlF_k{FD^ L EKٓAZ bT a:LJ3oJ(0pT2BI5MM}_ܬQ-@Ѻ~ё|G'<嬹ѷ+o\DcdXզvAb":(NҘvhvBBq=Z$!Sbɪ+d[LH!A%7`vCpd6(N5XRV_]ɰ_F#l[ xjbbHe1<뜀8(2":gzٌc{ ,leBDsG"@eI(8IF6Tesdeld0&Lg8&LΠZ}m9$§ }^%+v\!S<.4!;q|3.rRخr_夐vTN*1x)u5_zCj>lm{Y8aV=շ"󎦪˵ŵSK.;PV6Cc\sob A 1'1d+L }fY4{u$Q5t'Z,I .--Gʉutl.< WﮟZtT WsCCΡ%Ύ`MR3'Λ,_^vt;&w7w.%ᅨwo_ts~{ƂՏoU۟_y{Qv|U۹݇Ka'K*ץ0WrѸ>hg⦪md׻taJgѬVuЂS&(MsNWt^S^|wչy]:tݡ{d`uCW3;ygGY.uh aRyNOq=pF[:[45R5=+l\r&}!R _L'ˉ7I~v;\$M˟G 'Tݫ*ĵ٢C夽 !fsϭ27fZ^MZ,~~q}X \VjuY.eVWڎpnX!"F5ba 9_ Fg3o0Ċsz;qE;(y0IiSWea{I^iu ߁mC_YC9k_i %OBz6|"P;/3)XsqADz>Ry8- }O;Ji,– :v7/3t?/DZYF#30#2$Ȇ|<2υ|(!b㩼O\{E!+x26Mj8뾶CLl}K7\[),NAEK^{ A0O%!hR)G}U"&nTxh8< '>~C\y޺q^n'%8* ͟e/ JLz!l|P+8PV./KA=Ā,^>NAXeb{O,;ˏf} !TX+C(+Úo+A.|0B>g_s? Cxeehb͵Q3F@}m mjDlZ%zo3] 6"N!=Փq՘Ztp}eTL^NKI0VO`a j0ud5;H !dű=ŕKprBDs0ylB 4 mg] ? VzXYڨQVp&*= N:^dWAnCE}(0p4yWMzvKPAX~o0TyҴqnq:&腤ikKmKǃ;&J]+&Y#ž\&eÅV!:5TCRHUH)%TQ!ΤɈ,Sa82h{݅VA8BdVO5P)gDr9J v˱@\GD˔h !V Yp97O>"Z%6ZD_-RLԆAR3m TVAJ0N$ZT@ATH">0o _~Z팏BtZEW'a$2(qS"o痜OB:^,v_X$P*ΰc:Np37RLE F(94- JktG;6)^ެSpv1W|qj{?-Mhךtz803[FSA;w;'/QY+JU#` ;x1űC\eSlj )T4ɣ"0%[8Ā|-Z/$Ég ,l'pZ$!lXV>w$u@Vz.+X`UR0u$j._a6]G7%O4v_lLg&s\]5 $|jy8[}Si'WʚWZgNDD %\W$*JpVE5Pk .`$'΅ s4NjrׅA < %d͡~]FïذŔyC+UKG.䠷4b+/~*r 9 @]^9el}o! ˶~ 18H{vxCEZ@XX_^9~ُ8 U돋jC+K pb/'D 1`)ü7vpX@~ IRi͐D+K!LFfK7;O6+B(j]3-s ̤,[U5(̉b_֠ve i !v~\e$sk5bǍg,7kv:_y}âT;/?|^?:7>AZUtū"5knnrmѮI7/?^J$$EK&R Dx_ $/SQf¢d"2 w 3P(T1LSZ@ 056%J9II뢤$GK`6Q̥̱IRJ1Kv#`7 DjN5N'TQ"$c!D6{|ΰ)FC ٜo'3[j jk2dV* Di.-ќScK %[ci4aF#f-(_LL][VY­mA`R7wM:1y˥W*Mz_fAk(Hy|\~V x0--?D|ׇ;e SaE/0~{1Ht?ޏFlQSNg{On 57F0g8P+?L$QV8PNZ\!5ZZ_N *(/A9eBЧ%GK6&"$P|Kꋬ ӺI>iQ)hPv@U.*ײPazlk?kXE;)(5)Z+5EZN,>=Deyz,]X-?7WWzrAY0!Hc=WLDg""&='!X2t:P.ad/Gc90&O&W ` y Bza-` FXDIYPAq}r8i]rBg(:VDT 먕 xvn|'9xsݺa 7юf7;кXpHenY`#JrDEܿ/WjpY Pfi}|vyJ1 1ʤHAs%Ls@A I&&SEzS_Dc"2JɤHT*1DkiM%@* m&X寴7Jۀ hk[դS mNkM_ݻm{)X4}~(8%_^ $nIB]9~խdo^n/3fBTXXchf/)w)`?e`폿g;o ٥J2.gy/O4T3; \s| k,sdj(CԱd*p0\ OR΋rʁ[6ZCrkDmFc}ֆtwgF_K;Q!D%cyBie!N"x^1'.e"?$a:=A1dzwM&p( ndz׃Bͫӫ)ڦWq2W.JML` #D;Jh"(.d?^ %#ZGz)]d)l0,i3ě{ <{M iA<-8Ijp46){MBG Q2ASJ#Ua2,8,OnjOrlMڦ\BWcXgWԕC%gS!Bk eeJ2LQ5'/ +@ph xYMuIy+hڔ7-m9f%5;|)gQJ)Ɵ-PR:| yцFFp*~$E|YpN5{k%3I MdD$ʨ"<Pmdhd>)k裎<,,uF%( XP^6֠-5(MP^ig05~^r(X>*δO:WI A#>PnZ9 tjm_mQY3'S{jN I32Fqw󑂰ʥmğj3YS\y33nr0@l 2{DFa苁-N#/ )b:8x &ԗt XdLr*Rz&UC0oƓᗜT`B)|m-Ҩ d7Y$3pH K(p"TE:mngl KQ! X]R)4@qyW l+% 25EөV޺.N뚘hvK so( mQb71rY*HesQH$: a@ 1BThE4߿:JL3iYQ2li5BhzHojc O8FFxB$S%cbou>SHX/t5VLvl]6@)Zk*bH7M݆U6[Z $@貭%2Zmm>qрj23sITwDR{;hK+|!4ӏ.|^GDϛpʽ+MndRw/S=h^=Ak̪H[p$q!2_yu(Vvyh}ؐ'^bI-]=,?4B&~^6Ig3~ƾi%DQh d>F"`RErJ[ksP#?]D#a|S+~=[C`yJTRr =ܥ@1s{jhBNX.A}Yې[,ހ ț?XUDh5l"F rsç'4F*^$ Gh,rL:Z,Y}ESuޱ9\uS\d泼?r5{p*ir58a?0CSH!: g9(g- vtZ%3Yj_Ŀ̥QO_#8gr)JPȊ4#c-MP}e'֡$tEdH<EhriPvƐVkugz_gi, 3pr$هM Q"^Iv2ߢ$mKZOdlQUEJ+`Frثybxيx{Qu==EW83}020S ^kYnYaXIMW@8LRڹmoaI쓹=ĭwiQmw("ho1 -WQU"r̗q"qR$U*mau)Adb>Dx]LLV%c yǟF=f)=C IWc?WT㊓sC<04d #ժh3 kWj!#d75^dDq1L"GS)JW?GA`h) ;OW;;%%/%cM"*@@ђ9$Љ+Q3hkgFM,:Y#5)F99AbBEopYGAym. d^xԽ=s63)MiHZJ|{u/*! (;޲?'w^Ɂ0T^C~ڶXJwT/Jт5D7uC"Lo$&.JϽR4 ֠*nhCTKOݠ/lyLX{1hM:w}wCڛӫVtqC ikЮ_4o><~'-eE@cc>fooq~$=Xd+NZZG_fO]~dc~Nvz-%Ŧ}Tc/[-:g_O|7օ ~>G~-W-"Wa[GMZ7}8b:2&ry6VI"Q>vԳ$pň.A`Goߎ߾b4]-'؆0^ l[ͮ4JK\|IBevSW5*eJ}ωZBQS*mš%ow5n\^ܱJAFgW3\*4w꧚ hQvwZ=ڣuOUp N'{mvym}hJNR큓sN<9+8=:+xQ!ܾ&>ZzN( ŊK%u"B( ȉ$m|GJMw*(Ŋ5ߎd@P޹a:@l)%~[-=wv!\W6?,"^!w_{^d(g/+urh[nWwm\4z9`@RPl~cgh1XJ[-ouk b:܏;᪼{nԸͺM(RϞ-:gM=9-H >wk=GA1 +Qz&иմV^%tD1XC\";{XPy}|rAoz շep]`uQML'r0U򨠞 /̲#6]?zݻꌐKdž[}Hpz=XܺU.2 œp{ͳ:wqvcJjHM#.jwvv.*S) v2۹)+c]ݘUcjVI\[.~e> ( uT\N Rc6IȱzMB9)É`,j.X^@5C. $EL[He|+J;oTGZiJo2Y6jgM ",As4p^:K wyLGs?ގyނ;}GFВ6k{W(ByТJ\vgx9 P$u8yܴ;bݗsjfBy{gR F_,~2/s# aH:|܄_N']aLq%~)rNQpFSXffӘ9WSb'3\rCqFWT?f<[(C$fBw V}=4ɬ $Ml;/m О(Lp0C@c'ɔ Vg3&ob^YfoՉQ_]\>[Gqs~;NpA̪i»҄+RǵvG2 *\Reݺ󉶻ݗTQJe,S?$} WSʗɶ:p}Di_:kAWǘh@-,g`%4 \fue &7IylŸ̐2U-1߅Ӟw RwG0+5hC$DM̅1;.NT7^.%|v%Ɗk% 'K3dx|%;_l-Qt]o+m]xNs 0eb5}w|]<~E:Bt׊K!m$ mw1j-UKaFhLaՓҿw9;{K KaMBnFxkw!(k נ)t4N5^-N?MZ_ctKʫT^ `Z-'Z\@\9jf4?׻W7YA%'hs(<œ*2^7C¬7׷(wkF ^ E%N]ۚ8I$-b׸4PW=WhT5Lkب[>^U〿xHsyV|NIXRӧwy@>OK݇jfV:ޑŤj<'VJddR,"4(`ΓɆY;nCR(rzv0hQ׊RFjBJE4sDt&j ɇd%IZʥ5!Vb3XljQiІjX.PƠrY$dH];odq5e 1j!i>*/'v:Jv2dYe#|~r=KW?Qߞ]?t&w0|0XXƄx2A9;8gl^U6Q(qxB,Pf]RRƤp]$֦lJ+ч3&8uӘ|,>69P \a~G3\BKhs ˹ΉJ  $`$BiR'ID9{]ʡqN94)9иr.W!uk79P|V4s566\q Owƀ

U2&6lXg7g)hO/v>e;" "MRElM3Ii{wq^w? 2DC12p u"&3 BOïZoA;ROQ[Xӈ$;h>F]6@w(o$3.Ѻ5PqtN fsflJfRiH"A+_4J5kQ&zښ\g6-z4Jk1&kUo54krL|Q$բ\@۔&W3ɭ<~g߿owLFeVh]WHfɥ[!c zgfݟT5=|kTA%TY}Z HO[M}≕4hb]2bAC#ثheƝyvJ>_ W\!Z xkÁإAK#޺o>q4Z7shL )vKb=Ɲ4lyMK`X?[oPI uɞۢ~2 mV~nL緵;'FӆJXmͨT~YӧcEܑ~ح—/3oo^[3a;^> ϳ7] jdvV% 􍶕ePƉD[*ڵ˜n2;pWBGwg֘b8C o8X{V8$ug~uwbs`prp#@f`Xz(B ãz@E!;= AFeVe 5ME&jjhΎ̏áh?w^@+­F/n㟿{}?勛yqsH`iTxzz|7>΀иC!- r'?d~zOl7Hz}%ms$r>L"E ӷ䃉 鄢g{Վ\|^tdP7EG>M272AX[i($'*#hRhH+-r%|'hP Rv$hKhd)5>Ct$+Q !ˤDBR>ȅb*ߦ.+Yܟ, Z^Ϸ1y}r4 {-06&$iS)6 ޯ4\!cfi;-Xzd1H%'uQV)L. 8^=n~LCjVz+.g7ۯs vlZ0z/HB*?ǍUɮ'Ƹh4lUC*ˆlNJD'['a$lpYAllT$M.Ƣ"z @yer}yGlbn&Cs2;I-5y?bnkJQuIJZV;{~t3T?׳!~cGg~7Ob?? p ?}|ݛcT6x=o;͟|\!/.Jf2-&Dy7JQ2MGxփq\3tߩ:Dm2boюi8^Cl;+7͇Ͽ7U 9|.՘ tp7Ai߀b%fL(m69)kQ툽|26t4VfKJ,˥qŀ|2Zҧ m0㱜AIeEL#H^ !RفƘPSGlv0-q~v1 @-_4R5aI+r6+ Y{Q'j~CJ29R %A^ɭяuQ> ƗՇϕwՖeC)G.![-AW,t<;'YxuW«6#{wɳwJO&D)gTh+ $)Dz|mF`H']#IdD)7fa"k'&b!&Q<O2g?9| 4,o3X/7DBb^ꂖgK ,8#XS!I {{2pS)3/Eg ?rʀ7%} %H=N `;`&Ǯ4Yl@^7FR\V.Lht$ſRyIFz>bdEGQeuK0؅,ASycO)dG}-KD '$2[S-5ۢ'dYv܀!K.?@4!n$8Hlh}tk:t@M^?(5Ft}@3vXx놜E[%l8danlj(}m_z[ȱ/dU"KrUT:nwLJT}4a@I,LKn ѻtPehRUŘ@dpvPꐮp2^Zn‹zj%_h"F;8tkq/YKzHU^R eǻ}nAJM8 5b C+C 2!`^X%4^Z8uy;j!<뒔PZBTxʇ(sCޗR/MTn/&=~q㭏 @*,I8JB+G&A|aP B6EJҞ/>(c2F+ ^T[ a⵷E)QHA|:l3Y"%`UgՋfV_AvRtyyYTgi5#cKϹPS`vY""x{ܲ K]vrr qr$I6b̈d/s&;eA/n5=kƬsW$^60;`!Ǽ9y^3]7wzߋbR+K=¿[ף?v[[ Qkq4sg"C< pra~'I ҉1XU:QIV&mnxuv5d\.󴤔Ҍ85:-obK t0b'PGuxh#UǒA 0SivL/Z/U(FK 5+x a%D8I{xF]}]a1HeVU5M⣛+v?eH ^pW;#NJsV[x!WՆ]uV mjc{A rn门 \-ɪEiIQx]pdױemDWYquPAkh^D`T16#:RzmVt{׈N-%V=DZAbgXV Є:c!iNyb"[]cuk:9*?[ ٿ9<:ryјauK-}'!hu 6wF†;us?" rMW}&f`-'f]jl1¦uٓzݸ6;N*n,5aq'OvU^u٦NntIydI.8 $Jnˢ;6(qФ'O,;');g'nKCk,uYKWM˃'Wm4ksy f=^=x c+7ckKnFN=EvcCuujץ讫S7uuz@c0ԁN06 X3s=qtudq?Uu_)ŢRe*A'*KQOӎK׹v*_B BmstvQ{ző=^C@Nɥ{Diq{ͦ8y]Ir)xŌZ)&XSi EYr@wvT_[]$-1]D1ǚI`gbA+?t*DBae)ҩ?P],ՒI{BkxY<1Z=.lPzcCP!/Ke4nՋ8mr "IA99+l@( R5 *}в̂wm$_!yC}ټ]f/ >,fulyﯚ%L ac٣_UWWUב?9p1j ::]j<23e o΁K5,1X!0R ,fz$sO;:g-muV&|EL^/bz׋jz,t TI5W[BnyU 脫ֶ KґXJΝWk[U}ҺFyJjqqהwTi]r>|0@̒<dUL4TNr{tOFnDž^":?PSA%eZ#90Vr 8[R4( f rXŘ!;i-۾^]xg]Ա"h X $i6(/%6VOsKJm$)vT6 '0%r5\$Wh(\K#Q(i۴9.}`DbB'a]zb0cT̊hLO lVsW(/ͲT7ДwTдwZ+'} LU;WOK__I`Dp7W8ޙtrF(Gtik.m=H$m=Y8"%%+) Ŧ RKXPpFh lppb+C 8ʽTrWc [=h(DpFGA*b (n1!,HG-b:;@1.6g#93/X%{D6cS2uZ= ; %dGTVN=qT +GK$wT+k kD) TsIPTw]]iVݓoB6&?ӺK)Cŝz!;d5Y<wljE1W6Dc.4TKQ;祚Q$ͥGq̵d$gh/F-zLcn&}2dd_&%= 47\O5QR>!Ce;+3(N[\$|ckj`j4kJ{|HxxwMs,i,luj0EъIR}w~wRQcRpÃf+6ӽd!vN޷~2Q)0-d.OF՜kMYh4Y^C奱)$B^n9nW|pٺD JGTӮQࠨ})K*8^Sͩ;fct<(nP! vhqzģ|Rz2@40°8c^lu,P+U $B2pփWLp(7X  R* sq˱Z7h$5_\l6O*_6 Xzn)[Rf^#%ҞDŽDML8(G@ZXJQ=> N9R t?7L2  Jt8Jn:B&+]e,(FZg=cٶey(ьsЇVH4#N4 XqNNsC\zf㞗8])!=pfY)Cj9<[r 0fsV(VZjaUJA 8V`_ )au5j4JNRlxFb R@O ]S S :T{?1q#,:QXX0`*>s]:Y2NLKeT.F崌aP)pRqR")Y)C[InhI ~M;~};5G\(RXDI3v!`}zLr]9& Fɚ&s}" 2h wKe% Cc Ww޽&*1xƈ4T3xGqx۔rT)w͍|Yv? vSSk@鋳k3X43s8HkQY,=_~f7נ쎾]ʨXTseY>QG]vUb-fU)kƮ.hX+m(U sɨ#-e* c(Z[:QBe+#.h*O 0 n;rP e)2t q;RlLZ,~U Lk!EloU=h:?w cdžt TxEkC|`t5E5\y~Bc6&̀Z'<Q^n˘ Vi>;?KcQHK˳c1rws2Uq2hTr6Pw߽n7E,V RN—p^|7_F>W3:[p9p`UP0dA ; ^RT|bQڧN!QQ$6W4UL(6k*H%u Ɠ+)J0V|L}+ /JX՟_jz%uƂ%ȥ1^ϖxqpsV$#%/HEQ<_7jPe,xBk!FU"^8AJpZrVFq 4˜eAL4C4YV/Sc>3y1G2!)=u2*w587s q r`Va^_8F8Q|; Q'#虩?jvfT]7Iʞ֚)Az7c=0]TM$̋ )ow: [aOl)b tbVXS=zU‚jQpd X+'HhvAρr]WMitjSE=&UjAE%t8dlHjG'L MaQ/0R$UQ,cp dG#HFӪrm\zs5R#U튟&#۠('p4r R VڥI5SbZIR9ƸdD3miH4?A"@%DKMU/LRlf襘Ło<襼rhzg@^?# 9Ėw7^n@;Tuz2{ U}7R4 { @݉lw/U}qJt=`*~b5H$-$A"O?7T{]gϵ #љX X; 3G{`mHpwMɬJUnku;ϣ`\yrg??fNl]^N'VpLjW|v2 " DzYBi< v4V]_/"]$~~w?{'G5ACcKCE\b7iz\1|Fa7hT6N%I?i= ʓf3ң~W&ۭ4AJ(m/~=?P"%5j,ɸ P#cЕyCCFIl-pg`!4lgTBkW KMI#y@mOK5s7!R%R*Q [rֶ}@2*CQJoOd ܝ?ŖO@A" 1aSJ$KVvrʽ qZCjQ.\kDؿ%#@բ\p5b8PrF9Gc(n'$9T\{CJWfI$W%bF#Gwk^մo|9q^zk`<>,8G(,joL=}VvK`d>"kĮI|Ѡcڠ5vF]' x1rhu #odqBepX֧ p&zfn/Wm.>_7hO"K??!j`±*T*RS- J㜓iiT[eQnXA~3z jFz3Zi087U>oO\p~2FW^>f~,fZ=˭Y氐# cT%LX/?&b;2K4gzn+C4!:p1`) ^Dr4 !(~3j- &&)r-- f߆ռ~0 Xl+CV WO2&HWc02o?{㿟X+v| nhpǯ. R-l4~3P|ŏﯮ@ *%{t};L+xkm#GE^$~X$bϜyZ nv'qleR-JIMXfbHoWT3P> 3cB1a?ޚwَOqAZ' UKm{)J؞=w^\nMc,7F,3f4oUj õ }MJ u)1' vFDKe˯UbtL4n(+T+ KBiũ*KABQYơ̶P!3B h*U0Z7NjJ销TK6ޠJ԰DK͔,e1Zq!"Q x-E`U5&5N,jJ"XW ox).BaJЄɈ&Ќbg3Z~^z)x鋽\k^â+WŌlNv't03L؎jDU h-XGO{ p܂CKwQ03gu,VQ|cU _E-$pn^.(f%%*J{`LL*KoT3)JpVQv;hy| {!d,IM1j?-ׄ mx@9*4|c**:**zK`bh,ta #t>:TC6(Ϫxޞx];سpOJ,|hj|]wtdq(MEe|oY6TZ4u*naQPLhL{"ӪKWk6Y}k5j^V mլnnlCRP. ܿG;ةzN I j 縭 )J+RWd*(X])LA+bPFf º})g }^]cO3FqgT4;^+[;*ݡs%l-f(Ke6+<"_anL x5_kb psVPC\ySpofp5Gȕ A> IR?1C83y@ƗMD&N@rC]< 7H2͏uT묐hۇZ7p62WQILȣ<dzgLUx-\X&GZWղ\=j?wED{ѓۻU)u1$[xyvaRrۡoYտ]95:#$/}v\/;kͺ+}j\ZORHYUdnId!'nQ6!c"9xX |L'6^)λ3yzzr&eSf[,>Fwq__0ww=[ 9q)6wѻbc:mxHN̻{zr&zMl)_ٶŲu({l){!ލO=hO|E7qK2ړτ貾<9aj v0l=vv[TQ=e(#N9# ݄43gBKF_0Eyad!Iym=Gûzľߴ7-ӎ6Y=õ99B5>[J2) #C~e|<Ͷ{ip|zEWlE/?RD!59jNe)C4mY4v8XK Rii)Ő0; 9"`LN;°Gp8cD@D.zTȧe}sS6=~rpeoVNl@{#/uyjҲxQ$_>6WxVXxK< ϶d U ]jI]F5AE L"TE3ZepZRnZ8J/hT 'V&5njTZ ]oҺf#m:fQ%-uݙ+N\/|d:߾Pۻ B3H}*hz#[.-]d`%Y>DA/q, kVu5xϋsЌcJ!{W@{aku_|0̋?]sP57Å% Ju Ȧ,SSx3-UqZC 6S ZSJ`*gCOe#6) QIGǧZ8)bbZgߧ @4'4AS\4O xE;q{h_)LKy(0CcQS z޲>) Ҕ ZX`A;0T6,BAոeS`LkN}n : UZEF6L`l$B ϱPP\WLbJ^$hfvTh2I $jj֔F%5@F(GLR EZ4RlZ20*A@UU$)2-jM+\֦|:%G2I*JMQ ^Z30JcRJ% X**^A< 7;$O)Յ=3/V^Bx{|74_k{ >E0u ko.&qe0nno?_]Ot{7wn㷫*о߹[y &Up)q=z \(.&՗/}Ʊ\",i6̼HE˕֍yZ YdQ$jYrYib 2)Hc0hDjlf%e+V&lm&!1^`!-RԨ0 W@wf6yJ@"r&M8a֠uxk jgXЗb 8 H>23p";p^*smi҄q?w^N]Zz%MAnV\e]_zmҹb zb~=xyz[IThrٻ]59N.)JFӡJܚG6뽶f ($QbZFo}r pzJ~|s(70A vj3;*ZNr*gTF#ěymYa7Ϝ=4#.?"O(*Ew Rbčq[0åV9 Ì#vv;Xjc[ )F` gܽam38V4eHZXFS4gQӜ#*Ț3toOLZ:N*+"-=]DZhu&maL[/1\O uch,:,w!"$=|9MI|zpk#%iF)%̋é,%%iʳv\ح'ڎ"[3'%N=QPQguݎ 1FO6Bj@wbt* V:>³t 9q)vEqo|<60 L0$A{iY"f` %(`Wp8g#t ;$#ʱөOFHFw)|ٷw RXȉ6C -c!ƨvHio3LH0\%!bIi`Q+w߽i(8"U>nG&AU==е .3jZ,R3/Ibk*&\]H4qIwb 1C ߢ g!'BF[BM\Jcfe?(K`Aua\#LN%͔VH1[ }]m>mNf? y .-w|*bzc-jHN,}O^BG+BbCwEZ~ DK4`hl_iC0P-˳CYΡ'=,/>iW!7Xϋ-i_O~%i^]W2lf5ض:K\A"&FwAq@7 a b, k|"$Ga=G Sqj;:h+ %pEon;x=Yk2f;_..).oUӗ+SrvD9qύe,_,8rKʯ.J'VSdJDԮpWow!Rl?\v d1T aB'(@`"t7 aJL>e#p(QXW1TT.QU;(YCVUJ*#S> - ЪPD Y4HVP6B!qI.0VY""~jܔ] 1SK:U)mM=Z҆m+э=Zh{F'`GAmT%<(QԽ=b )#QNٱꂚ@ˁ2M^,D/&T]!Gʉ' pr/eTg*󇫁S==-ARH= ÀTbMkWU>Bho[kI+Eç̸BSca p{ty;z11Z#=.H#պCp4yCz:Wvβ^=Wsm(ֱxE2-%bv. W]LZ`yE 9FENst9gz8_!'C]e@O78e% }d+RvE&)yxf83) `Kv|ucTl$R9k:r>PAse *'{γa>RT)ЪC qc;kK<yZ~VGaX֣ kձY' g~ kJl}uG9 GkxRdR"O69t-:z">NnV`=?(^Y!Y"cQ'!7x<~k+Xq9PV멷ywqsuˡ̟7PQ'n!=!d֒ۋuH ,F~"Hߟrd;B[I++n@Vȱ;^hPShCK+^}F?Mןe!&?/蒱 ~%/p}[oq+[G/fc>⅕ʛ?;uɒ2!fJ0t`PSdslj+,3tf'8q''#v/I-t ֛7jYϪ:ucdGN+hl]T{yg) Q΁g3n-EBlN8kIkQYdMFʕP Zp+\7ƒFz^ƎwA5>x?{Gfz-jR PNކ6HuFϯ9׌'M)g\[KH5af^@'?$]܍b]=<|gߍ72X.׋Zǯ:BMΌ5}W?5n]x^1Lh=OZe)\ !O.w)*4N#77H& k"% ,8IqiP+\)zʿ$GTiZa3UMԈYVuϳF%Z?N~}*%S$ (qMh)]Sc.(Ry]o@0 a7y~i^Wق^c6ݹ] u<Q,CF-Oe/93"` :;%(5ZəMt-*Vuza#c0Yٔ+B}vS#(ؑ$2OIQJ1`e-g!5n(I,\rQu4'ICyHI1Z'JY&T 0jDn}"3 : hm9ȸ#X#G-S+;mUX5+Q*w4`VM֠8XNY(0pa!:R+UgL&(}+>nvfbHeRiS1WKwJX%Rg_n+ bP̈́ J=T5tt)Z1SbU{jlՍ9bV˒H@a'm(~rRB/l=-J,,rl7$OSٶ-ڹ}J$\)X̚$KK <0L~x;훏oKa~;&T< "<, \O‚4]2s<~!Ŝ,h^ׯn;<7uC?|g0C&1 %C~$S>?`w.td<%f?L&|!pmErH ,\=-L[4m\2, F=I4Y͊ JJA%E$Tܣ$Hiu*AH:n Rs+m.$#c[5R*cA/BIlCns`!-};ׅU$'G)zAA^V.JYȘ 3!vIE+l}:ՀfK$q{9dlGy(؈lR ϴPi★de4&6wIIPMnqμcEw۠H[8fK+I=[QfK);#@")Q=G吢r@%BWPf Zn]IkÒ/g1V6췈-)h,gn3K,綅їmbfIE jko2` VKz | Jqؕ ƚ$o7kJv A5Ӛ@7$aY'` #!]!W Jܩ?f:Tl^%jȦC_%+*%R#!qoPx3Hrv VcH5YQ6d L?O֗!T0 摨ylW<B:"t 薴d]agH-9֫<{V^RCU b1βLQVޑ$ޡ$\sI(Fdp% !'&g YKr75Tz۬j G w64pN`g%݉4ԴZԁK+S 5-9עW/PWX*- -{`C^fϪT#F ̑\&<%Ep~gZe[w@?ssM]>/KgDzFq/Z׳I8BqA"%_A׷cRQр_>|Ao=pxV tc~kQc֠$UC%ֱ_WgJ|ǿ[&Շz1ª }/9KΏje Oyz C]Anv -o S s-3h_]}l? 77t),Kk.x{d]d?\޼O<,Fn.oPB<\>>ݽk|*#ӪɮεyRW6yz̛<2&<Z=8sG% dsl9c0Z| ,=EΚ\M8_]i^O~p կ[pFpH)D 9gW6}G] ݘ(rTZYzKȥLd2$D9P[{)g I[@J8<&Q _@(,N,?W]SO ~TSjSYQNx#,B%LG( y"L]^L׀LP"s_<(,zӹ4^&WNJ8x1w1`q|ݴ|#˱EyZruQy`#HALi{(]]&)~q@oVv"=[.(%Gf {=o)o9Jlrj?_{ 9$Xߩ= _Zeû2ߗ[h_im6ڛˋdt܀F1 Ǥs恄♃:[>m<۠,\0: hk( @ 7 WV+}~Z| 1ZQ̬S6.-w.FnvvL]&m,8z͒\9xRzlkr)g,X[L膙4T:ۡR!)42f+oΑiEޖ ][kep nY# LHOV-@xIJL5g+KƵDΊ4#bƜwqH4nwO6_ A9,&|AdܷnOf~EuVٔ(Չ qjTX,"QYA$=d*,*h`pB+. E ؍q +z\YJЂ(o]0X PK%+ifXfeh3#iG@`DuoQmඐzX^ϹlBM#Am(eiueG0(O&8(+LXRDOSfKlH)/nOnq>~ğ\>y60ggvz}vaoa8Ny*EPt׋* QcY3e a^c@-):bj?(n7QqJ@8囵b´˷ a*e!_);F+; ޭ:1b*zt)J" =/ kbiiU]ŃT7АƸ.J.X9XsqǽFN&y&aSjo \ۻ|X&FE$*2Yۄ|6,zer( zQo[c7Q`#ރC}v(ve5lRU{tSN7 hhZmz=]d\S; !9^l6QH#7<Ȩinc-cA!yʗȜ g_%ĸ2{*l˺ţ'gͥj/`{VUZ|NT"zX/;7g{\n+,^e5.4㡼H 8r_v7t-'T[X NR'MbQp\Zja3JJsI*SXעjr[W~\c 8ݐl4҄D^:UCŏS_NI9/fj4c_L H"=]h) &V#^jBDͣu)63AzayNtNFvft*h)eSjJ9/GPn 2qm96bxC{?|#5/vV=nCl%_"á1"p ޠO{1^)'X WRիbʴsbbʤSKbtTDޡ`[)RF*Դ6-g&C*RpDmbf QZ7#}qKo(1pdi-n.0Ph"=Ll8#kU4Z!:=R!34wW'*YHˑdb-8y pyn1Ԋa%ty3H8(BɆpc 9-|0Z#)WR3RuR$sB+^Y|[Qx<]%~9O7fgá/rIronm:_N֩N[Gw0NVIuȉZU!W~Đ M4˦T۞fk%`x\vʣ6n9 K8;Q[v>;4T <Իl+I֚|69DtʯQh/*Z&9@3EeMz/wqCeyH&z#{r?^i9bϝ9LQ*P:??3hPtW>a"5Bm3D*=4T.)kg4ICp-H4Ta4Thk 5jT*4׎b qY/F1|aHVrY26΢mU~%SSֆ /Td% FɒB-49%0^聢Pb{"Nq⸈E}X@B#ɨgpH%Uɼua΄Q襉m.22hmVXNDsw;^V?ǹuu$6<vkN~5=[=,LRճO2^^!N;~#VO|_^+<:/5S7lkUTF@f%08Q Q%X׊shӐ\`\3%;킁9` #qZkYmDV6,2HKƐaib7C([ɐ Xc*wŠ&bEO]b$ib'1H  UKP"79$:U` jFTqM%=io0Xڱ f5v dHzJ]Ixw~O`MB=ɕM;avN)UGDUl"K)lͽG/Z\n }y3~.`J]"Glj]];^^X0<%XqB~Yc7ψ+•s+VH#7\{u^|]&xU2 jõTնt*s]ňL%(`}B*pj%]Ai\Q:ydEJZ=9|~~wXyw'FSs%%;?[[L V[j  ?t o=1̴:&lL嘱^&Njubc`Юb Ť 0By;ⶐ_6yuy$=sf~io̝ ^^>qSs5YTzXţws+B1j:Sf'V z0سB%3չpzO1,b9SUhQ1ͦ|{oxɊO瑠ew{Z-77̔~%ׁ%˂AY(Bh+ݲ[nPTt0(!u "8S`87B& c8*ťUo@Kugc]QwZp$Vrn> Sycc/{ x05d & [^Ћ!e?bH8RZ5ѨRΩ,hxbކ Xi)p2 g$:?ڡgfcݾXam2IGY)~|i1^z픇^5L2X,dP $u9 .n{{A6ytk1VTKvSjkHA a3o5b0|Hم|&ަ(72pbxۃwGҵxziKlodS4xr8/߂'RBMs 2EC93?uC.櫓KL~hɨf`+Ta L^xjWȀXs3\ˌ>:#%Gj$ ՃMK c ORS̏G\j'$:`q\z6/J8c#JU_Ȟ^-0d{):܁y_9^"z EGO zz/ RQU))4 &/1Ο6 q8.0>?X J0aR0sAKzʙP:̗XxI}POۛ#`NA{f~~YASBKdz|g#\=|`onnن2QLy1 \s|Bڭt(ŚN!$iYXaR!N*eS )Jd ͈(ֆiV%\C{N,>h v>4)VԀ%Q4(J+1̉f03}40"r{I$![@al*/Sc4NQAHtrJрo0-xi+)!)cp^ȒZRD[nT4zX8!mt5#YB/9#bOX`q'/ JBRv|率{3áeq8]}՗xcXM] #Cn$vZxA Xj(S< \n).x"UqQ Pre!L+~ &Wz\jAT-@BX52 ]ηmATC **XY!] ij:&6- <5J$ e1DJ=@VDR{y9g(X1a]"(.L|ķ;qAD73 r.p0"|&KLaXnf4JA/7ئx,&lT4HR[_*倡&3uMDa&\%'=kdK.5i4[JJ*fXX2C!}L ã7s 4BZ`[0g_1-"ěJ * tHjڸ<iIK!gCn\P>孥3lK˲҆hf-:@lٿAV,k2no";9*Kalks->\"yC*R-Z}.C325GߚdZ>;h+TF9dHy6˧Whù.јuhҶ->4K$ڦ{:{PXa‡3=>sqxlA6b!v8"@^k7%*Ia[1z @a8t ) d4. @%7 % `v-'O&뀷oqDQƑ%c㥠V|@+̴Bd3D(pBrh! 0Jhae$_܃d R1>(3'-D1jSB9W:M%6h*rO3:($ZB¯:oVVWdsFeq[(eVeZ$h" #Ĕ+i b0|bDcʼnO,%* y0w#LälCLb`{=.[P>˻DCAl(pJZ(v *)@x띋 aCȵ^AN"6Z?Ӱ-zLf64䅫hN=_ZX߹Oxݖ{KB$ !YC@d,j8"B ڍ|H`CjRx½NG1Zf>$(D@?.JQex@?#\F?p1p̮}L? y*ZS~شn4z~ȊSd)lQũy$Sͮ8Ղp-)BJxxZ"bC<x_.s۬}Y " ߯az$ 5%tkC,Bݧ-<Ta'`-I# ٨*it1qo!ʰ]Ԋ.jzDhEg2/:Mm/K9gol+(@p`mx $'l#{𝍴a,VU$On:QWJʤ챶ܽ=h\ي`*,E 9pI ibk]}Vkz)9':Gx1˿lEJMJ쑞B'6F#/ʹhF8X;ET0YWboZh9ڙiz~'( !ꬪ<\ܜ?7YXJI1WC_Q 8#ݓOfՃƦг x~mG|,*ve~͊O\?egqQƱб3{xm.݅ ejmL艠-CZ}C}sK ︆i7Yvw";WElm߽ݓ"H0^j} E˞"L@()z<}Cmf}XJhǫǓ9ѯ>M ^#m|R}ݯ{^=$r B&(,;MW+iQ:{>JǼ*~@\qeXL7Jo0҇?SH /Gr9HiC) pcn؟ _oG1"}Z騴ףU*ȓD Q=k`=M30ōﭹ ݿl&!G唲ט/缡0J$<Ŗan5ƂeBjG 1 .A=b$*Pi 7[ xTA(`S[B$Ԇ Y8, Q)uC[ aqZai8"f` Έ@ CP$"Zw1vb^ΧS b::f@[#{\kֽx7h!炠/ŻϒfWGi @JDѯ= t6_aUOF^IsfWETrs$<ܐm=┄)ۣ7Hb }KqRAN/hS0MS:u/ۆ^`VmV?Z2 B$SLicGq0xRPWۣ(Ak9goVrJ1q )qG?01%"k h6{ aTaK%-I(֑qq 1QPG)b'k5 W-5r-+!џlls4\u{Kqph{[Hd/c?f[ 1.}HwW˄ܿȉI"w,^j8f0uudY`@N/L6٣\N0`qrZ l~8]2y8-_=yEs~򆶡M)n퓅oxx<h) #~.bzHKPgH?oo'IwMuӻI'9ytgt|g_9hL74[- \G AySw7_&R-9bwe^֙%/. ,q{'mM~8,_㳿oS]݅' \آ5[m*mw̨Dww-N}ϑh qACOW\mI?_V/͜Nt3SpvV N/.A@NT0mA{[}"C3 aQXeg@c=)~:3;Q (&",5c1zfq c-(1_tXKy#+k S: 2!3 5U `6tCI9 ȼ%PMI'͜>i YV$H3CHc \EAy#XI DI&noƌJ]/sEu|.۟~Cmwi Nݝ/N0AmC=9UvǓ2y7_XO.:2ٶM9@w&A}{ZLDx*5 L›Ff;1˘1ruUMoΦ&p3Y֍ތfcV5nf¿ݕMn1l&Q\cYe<4r!^N[C퍛-~惉&c5.m_Ш=l~]ާ2WݜuUoKYRz*[ҐE:,%a݄Lrкb:m4n&Ign͵nmh W p:hҧM"[)9SFviںȴnmh W[qq{bc.$ɿrȹPja Jo.)h\҂p[`BVLQ9BBzIe3d'܇E$.xh6yL <6P0< U!f65AKrԩ(F`hDzm rBp୵Zbyʂ2% A(OtC5w79+@ ''qZX;1_/2"~Y^.,鑽[4 < ٻ6r#].5WRsuuZ'_ri)%Hy߯1ɑ0ph{J&|hw&P<?~{y/}'`OwQ!R5k ?gӛs`狕|0$>UBߙ+ܮ'Z`Z'm.}1ܪs~2 w]v(铤uRp*+GusQ|e2pDbK~4u}g0tMCQixŸgFOc N@ثi6h=a$5(jӛhL*M&/B['z)z3 &- aNsY*iՑ,^_~6k{2,3_*:b4sڀuTZae"vD1i/ 4#BãR 8섖]RqT5_/f(Yl2M` zmk@ Wt$|]>W{mHj'$^riuHBb8hXj%Pa54iwk4Ng2S` '^#؅@| AJ)f2TTޛ97„夁& L棥T xu{MKϤO翋U k>A(B;dJ:2LUu(NEcB9\\du䙾| NIcS K5[Pd7-`5sK8 %"J*ǂa΄QfAӆBe/RKJi$e[$2%a {n6@ K9-a[B#+.]S5e+<{N;4G^ǝ&; qZjΙ!2* 'ksSy~9)(B+Z)x% vN6销 xQ励0{82)'F~ZhN˼;+B7&3x+!Yx1eț'XwWL21Al5a;8ϧDutKsZbᇮ~[ʄmAʄg5n+^\}#f+/=رvGW\^˾Tv1Q՘' ѲU{=2<%H`bm5cE) I]C8tx(ZסDKlJp>~ۻQF\ǼYAM1:]<ڲ tx(Ʉ={D)%b 75M#ܛǎ5D?ؖ@„bL>(8`0clo%JgtE(w#DB,~D?};D%CJTa1Uk]JPFQEOQ _q͕=re3n*Ng)f*wvvU"H !gjw>9*XݘY %{ c`ʧmQ1x2+S,HJyv~LPؼ$4>"M Aς|o~$Ɗhq7,[Ƙ(rgIcϨ }$,Mb~3{{:\r3Jj9zU^C?X T[YQ<_Rp6Sц-oa1k|1I(ZI/0HroU:xqa&=-݀+IQ.ƚg.u+ZELbLIt#|Aó֑9#}ogAA42d6kH B!(fB &Sz;gW>Ts@;i7ڒ' UDP1HGI |ׇ`f_-/@Qn"b45 gɌA͚sR!}G"wVJi5Jk3A0m7 +0?!UKp}Jb4! ^F5q))D*!bEOɮ:a %08*0aC(PN2) C /-MV`d0J~vRFJȮ$%e/^̨n1/S8¶w)u,hB.Bkf =mZcv0}}͝Q¤#/92:9gARV)Yicv1o0{k<2Œ!1}>,/^'Kx>+%%L7aqZ{|?>@L+fv E˳+@dkX'p?y 5?|9K^oWDtoBZW8@̈JFXe#왃68it":YM}c "3_NORCR}kWYzjU%$N&Y@~myxSI tE;_zp8L9rRrs(]Y'hDo@>DMtgʎ/(lҐMNRћ*"*ƣ RiAMnm"{ 9+ˎ8qkL\43Ƭ#[bubE:e[r3 8,SVL NT 8 :r,IRf4F>{Fοlx4(;㚛|krˬU?ߥWv5Bb>vake|"E'L?$g!9ə~h:UۊGbWzl )с ⌗ p؀WtVϑ y_ ]"ɮ)|@Vt -eE /ڵ/|>H.9ՎY̫Y{VЁ#]d-hЄFT j% (zv' _ljnPhKO3­)k S0pI_HQ$ReJ(-eNFoWyb=Jb7Ofcziu4f!cGSH qRj~9tlѣC>'&#A(oDרbR**UInFBPRDḌט -2#C h"51l>zpvJU}o4hziܟx\1\ՆӢ91`st8͋;ˍ9fSmw5A^:@d8AjHقg0ir?v:ePK= G{q\ag7HNVFeʛ9'=QX#JK/[ vŵ]YsK/qZ ѢU[F)o @ʣFi(o2y+k51(Sj>o磫q# ' yp~dO8?'$J8̮)8"[1n˵A y 4(\ A)\gkXeq-!0 2ǎG<+7m4~ `H=1[' Y?~+A~)e"Hq ܆ &J+Ř΢-@Y!:J\,Do"hr]ͦNn/RKCːҐRPIȯBp\tN:%*E tL.RjyH!Ԃj>y}#us:XO);k*UY ![SJwnИ&s_iuww0;7S_1Hl[%v9mē/JXJU߆yW*EӬd|zwat=-Kxu8k920[N,N5 U.3zDlJo;MF #zR rLt.1 TѼ[z>,+7уmQXը~ve%q?e.(׬X:͜6`4v+D`>s4Ќ8Q{b)'D_ oC%dHO޳Y~Ay=RKoWK󰸺xoϋ׍ϒ#fvpwg;/Mp~C|JB A)-sYibΛY@G#.n]'ـ9Wk},?i8e QE*&[XVɩY62|@'эW( ^)b*ӏ/s}_̧Yv>e%)2=o;E$EoNZ \k"̦*%oJjܖg$thbZڴ][-E-iUQǁSAՐƪ/;%ٗGW~1E:wn;mvZ0xg?R :i[,tpwv{wUZ5l/gr5eJ:>\]w.X}}G7RmB+O_jvhċe6( .z0(8ұŅQvA"/2 y $ZHn>s ~1 DT9S 6GPf%B;Cv. c0L$^R:꤯u'5=%΁b+ar GA0αQY핌TC!X턒D e4K"#Vޕ` 9b(bϭ𳰔)L0AJJuX )aJ;B+RҔ=?1_`yOb)dKsEKͅ'bۡdz-`{ChY/,1ViTjU1pv0 ?{۸_$aw&b/`&~ 1Ec{%yEd(I6)#`у"U]]]t(&c9# Q.3ͤJP\`,Q:I9Պ | a2S'7ØfwSWiZ7 RR˱1DbDa`dn[YrSbnVٗ:V?wy/AWQy! H, wep=xӸ8?vk# bH*!qRQ}*AZ~0J2z(J5?h4W]A;BW+Ojzl/ws%JYGvRA]ʎPMZw=DW\s-~Nn: $cՓM*1' |T,M%R,1ȧ M5*ej%u& Z-v*Rֿ'omڄh("I g6YR&KQ 蛧6 OA0Ff s0_@B ]|zcq ߌD] u$,Q;spKgI3OZ6wVhE` Itq >6$k7Blq;8/ Zy-0™`/ -\<+|e-0QB%,Q/PFY%ۀz[nqP*ɬ7#2o^{J& SE|wvH\=\ـNon]m냙vqk%i4 e9𑅃FO|s,=^Y`,S90\1#\Z0 *͉X4jtT-|,3m49`2Ԉ0At5*VL,`(=L+[jyѭ_n͘""}eƍ172OaTh-jqFX`)uW[ط㻋sAې+2ƉE` g*}odU~{CT V}_풢PK5щ$[B@PeMzw[Ƥ}GtnEokO~,׾,vEp !"̆h^W*hH1kf02@<[7xm8nnmhnEmػ7sQv 3yZA-WLQ ^xkk|vܞ{XƳ{5X:_X1 V^Z%N8*+DsEcO.Z0/֦Ñc&`"IVIM۲Ik)3rlyoxWQ:j -b`ඔu o圍3$%ǚbpM 61v3n}}Ac9(G ly;l6g0Osn&(\XpLʗ } E9*4*[t`kAl1?v ,3A)BK錧,Pb@17(e] TiEqohx|u{1huwBxy!;MVXT{1hP~qj/ ]5{/WiFq*(}iFSK,zܢ:R_zfE0:=syu &#Qzf $F=%UTu3c]:bl}rj1FЈr}iML.Q@ K%H^Su\KN=^>wf#gҐ-[z,`EEn&HZMksX3r$MHо `4uFlicLN[{ .e|@>(-ԋ:n[;){ hq) H{g=ǜ,`"$\"/'Yn8"{sDkFh].dn`J@x=QzrZu)ai./ ͉V*|!*ͩ`dy $dtPmeA9N 7 ⛢S.+~1 [Az9B`Agk\;7/>K` <-EVˊa(z? bwᓍdjZtWu$Rћ|39o>~teILV9^<_I#($`E V\)6qFRR˚a0($@MAR9V nYg֌)=0<$'dQW%H ƨ&McUnά:*ϣ7(}ϿU OKW.ao5g[_lC,e bCVpnxɴ 2xָ 9#K&Qܠwy}@`CITkug rO""dxrԡ)Wbapv7|XjՑoFd`dD1,+ؐ[=GlirCW6OsDbL 8۝ݼJ ϭA%%F`!sVn-7*a.0dsS1:py* Σ_%D]j$TC6kI5uGL52IŭT}TctXU.]ߦLe B#dOv>ȩ||!;Żk i $/vug %P]r(h~TgGbB>X=F4%5y?W;_eyI z(DP%fܓ~tIA?Yb(=JHVL(fH`ј\0b/oNjvLg^(-F-Qіo$^(^( $ V2ݻT+[[ bD3hSEPoڭ@c[ ELa*rў\NpIWk Cנhʭ4GK^5ÖQghMci$G``FƇ uXj|)YF2:ʐC9^[m%yFoRG̍<&$CykC2fб&sJ %-q 9DCɀ}eSC_>ՈߌbÞghTRԊpP% m9lŞ#Aߧlu(+ K bt7BĞo)Bz'jӼ_aea~t7.^1. lv>W>ndTճ[ ^ԃ;}n1wH1x,H<Έ >  jh5_ KNS:'t}}_S`k5Z za $'/rhݕqr@d_ /pr+ÒI:zۓJW^ˣ^"JdJ^Nf#Jyq{[68 :F{2i9 -cqO.r~Kbf&/E\/un$Q:H9ƚќXa( ' 5BکNZI_`" JW"0r\5tXvQ4"L2[K(Hw=Z%yO.KA e%A".'.ҝ }HɋlK6e-A,琗]4vC(H`[&FVV1A(`JX8hOC" }Ydq)K( 7wf+^Mmѷzgb@Q?ݷz`:/wSxHgf/yB3^n1罀Eytb|?O0>I7/]XR(䂈 *(\B: ~_N?|xԶH]Cvk;Vm2 'o(< t0v6 7M,7픈1W0bUQ4Y 54{c =ѰXI?H! ($'\(F'}YͧFotFot|@IXu[(`Dܗ0I\r}[sJK1nCSwƇO:Hy ebhB d, ;.2){g$ EX# V|EXU&)h-1Kn\)$>Am;-6a\XkWuW:XqZ<9)h+DrKۅ%& d^Hۂn%06h!F$:j: mÖ3d dL- [0v[0 :*n#9 ^@Tm‡ۂq Z*[ӭxxI8tز~/3`$n{> N8`68{pL΀ l`ЗLٔ-m6xE;M9seq*&Lb@1Q2(kUvL$}\T&[:V_mܑy<gC !ިsHm;j1shu”TrAL@' SBH$6"˄1I 1H4wZm X Y6u$IıHp"S9nCM Wk1&S*9m5`rޤ FK6si3j1"rVHe7Xv$RYlU컉`apj e{z=ff/$ܳ:6.rQk&]yb4!]!<.{EMP.Dx1䪡\qA,%Lx{ cqt+[/V qX@ Q+wH v,hK#UONOΘI % Ͱ?56l{;\|?t_ jpO֡G|,x76˨ٙ<͕IgT6^[_ 魚??X3@MCnf3Eɶg2rYhDpf)d֗RVwתC$4OX^T۷J*xs d~ic:jJm9o0턀"WfylF$aP>jkߋUL`qe@`UD] `p@U$9>vQ7 evle!vYxoJ"B)J|QhGvGJy@_xBXV] /N-VTr_j}RJ ȕ*;ӺjWHsLx@l WMՖPsG^,S<)@qt| 9%gR)LOn%(0lgoN:;Ɍmꍜa=(mn<ydv5>=h=0+0n^ީynRg۱[O|lhDXd_gdɕe$BqQ$1瑁 1Vhg_pƩ %ln1Uq<1Iha@} !,P> s} 1|b7Qy BWJ>L fX#AU$!P1$6bRn~x};z:n\*Kccdd y"Hb t N}RWO7o<26sd|×k汄 GFMDDPj1%Ie`Ca9E$TɈ8xx+0#` p<e'urZRɲ3(K j>h0dD[o^|3Zex9^rħ78A@qѐ 8Io*KwfOt$}8je\2s(=dy.Į`!hƟF6P7ohll_54uw_~ V}OrB0B"=QG?dL/`Wv:T/ۖѰEޢYc>W_26҉)!_.>.:餝DGރKʰsCee{4TBDT q"?rTB7]TJCz#i㈷sРD z|/& 3Q FC)wf I(~=9b4ZQu7N-9hcn~[ PY||%f^g>Vw rwg:,8C}܅UɩG<1|\ DWBACqKV uBPq \a׮B6SSV9EҝXx$3* c&KY/^`p?07ݭz|k2Ӥ=yRN?SHllL9nKڜï""\֓:@*sXjj :xٓXFlBXY ȖJcgڵ=%}kě*)3 D1RlC.6b"R@`-b"EɄ @:T|BJ\;+Pk+@D`d`$Nzl">3)L٘AxPHlt !`3+H}bsyjZmIfW<*VOO/j䌠h&ͭLmപY Jyb_6=Nݿz{l60zqz,P[LE=$?# (@% N-bmV|Ө k9㫦 wHze`ήIpϣ≝y0%8$z q^b6B oA򻟿}yFBa)F$&NDC KaiD.O>yB쎐(LweT@FHL!sa]3HtIA'?68鐇"-È(L2yed @ϰYmI/IHymaq*EVe+y啑莌_x ڠIPX8ڨ8 Я k6ΰ^T*f.R(墥{ϑ5h0QuvA7w+cFJFc@@SZ1&6Pˆ(#acIB$hkWY'-? GAq4qM3/(nxډMS•S$Q?/n[ ,6n|+r*;Spɸ.ЀaK4nAg/ <^Bu/ȶ0fmfW~@xgٸ]X#6Tol UAq[۽sU.04\{;nRv<C DzB_7Y'ϽIe4ߠ>ܣ9u P8anFQ;y88%kY 2}f[ 7+~LvU6E@TKJ^WoKt%זbNrs$.D]So^áwrrsi˿|-?!m~7W,+榅Mhn+GšF[ir8O⁐6tDnCS`巓 ._eX {E]1K33*y1鍧gߕ=3\̻=T _6}M_yyB/zQAB]W2)>۹n@JP+aL Vf1А[@HAbaB,j +dAy9aWJ!sUWr dޗNzeyIo?KUs:p(MܨWjɫ'VZ}ʫt0E1E $HC5r$Cֱվ"$"42&1>{㳫ӷj)6[!p;,^>9v-A>~/]:]ZU߈z3=Lx` `tp5D h$"$vs%J4g؈h 0R3k˭`GPKU[Z0}-zW1r|Eux0Y?zsCA})(0U#gp o1s:9 )jı?{㶭z{#~(iQ$SERxmג$aq5- 3 g+'a̺UB @!ME"VɛDf [uDNr%W2r9Ub#FBACaJn07ШPѲxt׏Im.ƭp*+E; =!+9ew:^dr,2Ͽ)w8%RFW܀Lýtޏ';#p >9tAgO;c'?/-ߙ.zLxO<(̭! r:l0 5|TY>\f(NZ*άaћϧ:@H{ԕAԹ4o$ެ$CBEoƎY6O1mq9$"[^3i8eL*ь.N.fu1,?f )Ua5bO-j Y@A9Rq!2SaAf97qI)Vӱϐ c~L˿{Q>|5'QC%9IP+;Q|\9_lP:ن -*48kܩ](xsPUx /*$VQtPОb~Lbg01R5d'Ma d;XJ^C?ϰ#g~<%pgo ;(z,{OYibϯNQ~B/鯜'=M&Jx=Š[6gTGYQ[~Ϥ'Zuzt jʬCCh-lhG5]]s{:6˳2^U9eP Ƶ'ZV7!!߹.Sa&Q le o^Y@km6ݩIʎЊ7J.MK*AfEJFV.3=t<"H=kdm%k,5YT:#IuVH-Iĉ⒁Z47,;нi|}{|:v_)o 0 D0J dlEW%Vc_Y36pb\5VP cIY8xkRou 9m(2ՔV3 `2aV#ޣXR KNʧ+ň}b.~!5J1x; DV1նָ:&ͥS JYK;]]/{١R&eMERt=U@[!ُ.ۺ7&-wM0m[q\kih_CG9@j8D=^»hMkx̿k _m$ET/d +wo&W > eK+ xgC_=#h WP0W|cQ_7-%#MFZ3(_C BmmH eAZC琦Dzx1E_#S1 ] 0D=.ݢWBrG/+[2Viy4U/V"l0>? ,?A0ߎGշXT廷'KN AzHI $*ZUf. [{ s%$B,#D( KC" 1S<ZH.Nk/Z2z)L\dԄj!Ԁc%&dI2 C*Uh6ah4nh$dQq HwC>&zWvP4ITACh.d9DD:ԆPBYÕhlp`]?GFKe7\i;pc2tpi gh!qA $N<'ou"H_R*P)<"K+DƘP (0~0ڛtfwpե, 2,jG;ʽQ0|Eю0Tcv#L:$ڋN'[Br E5.e @0&a hMDDn VO]{[{ߖd|_Ҏ 0C/ 0;ޥ8s(IU_(ytl $ckl¥ ga)j>5`4SOMg.IiL\?_?BX#&T8W{/M5ZC:oDJ!VX6V9(jʗ m%~_MJ|m\8kke\-8Dj~A6u5X}yNeGWď". ~/! CU9h/[P&[ B#6<{mq^+]J@4R7{bN'yiz֏MS=`Ӧ *>-Qڝ~qpG{H(8u߽gMc^\yyB>D~HKS!g&qSK]3N}_ou"QX`*B hI"q2 p  +B@A;KvGlRٺ S :/ ^!Ak01Dc4 Y ar1(R*"B1]ZI!HdqDRVa#eF&dJ%S)!kh qrwث cpq 7?J'ENXG>۱vD ˆ3zʃ"5ր/I &@)&^ZpgZq#N͎^>^_:Y|Uh^^?> @D=XPwl~/TRx,wC) ?^8yibtd{ATG 5䌫|u釮pPFI"SD>{B:kY$ K0uQJ6K;TJt l")3Z0sB'LlU-#ĝ$<2J)B# 1Vis;HN䞏 bԹxkޤbT2_塕V Tyg?K#Ja|}bܻTRXJ%Ȕ.72qAmJrʽrH%';}U@ϪܐC"|?݆(^fs#"|h( G ôTSUᔟ=c;>"JK xœRCwkgLXb'0}S*NC%^Jk#w\y>yüÀC Ζyb\Z>b58gK<1*8kj.j[wX'~*0 价|`9lӣAHZn4Y܊PNzNFvX)ĊrkcO-丼XE!E$HC5]Q^I [( DIBaJDN2rk32we`MUBш` 5D# 1(VB9FDDsl ETS"yҺP<#ujjss)"D`O73Vuv>gk7)&? Xbl=O1Ş[|ggr?fjWTR)+_;Y*Gׂ4ms&$.+ګm%dғ6ɱ2wmHBs%S~n >h諸h8tmk-ݚ\DO)!%rAc5lep 㒓7Pr m?C_l2&,sk>d?$&ߟw6m+Jl&1 "?ufУBs ֩^ ؇'nܒ.3x;2T'Ov]V\3XQ.fYAX!`ڬsdˡ"ݻmZ`k .yK+<=Og$0e2W-e'4ij""ryA!#"toG;E߼o4W~Jɣџ&nQڧY-pVmd#C EOLH0!ӟ+㖿eoYj]sX፷ 2ٻHn#%Cf,/r ΗJ+E1GΌVό=Pk5akg{+֣Yo4Q>s7~d:L)g3^6P%iFIuCB :}eƚbNǣaY'Ł IfdyI բ A.Ж K%p@1LZb[*dYCL(,1L0ldL@ieDsR'!(2U%R 5s[ÍV^͂w)2PJ0iHQ7[ÄhL庠';H<1juʶ a%q)4 }(4Fb$*%xE%MZyzKla)yY1 ߻;@!d ,K!+]l K>cW+Ne>u2*dwE"G2FrHx!c (CfE|fٹ=Pv;jX]F{XBH4ݺ_f;CҐQ6]D6woϔ\jIndc]7h i,t41sRr K6o?‚Xp#㘝F[& 4?*=3G9~$!Y%d^nY->/iF篂Y.l.KHN2QܒN3j3G:E~/AD\vW{^~BvD;k`(nV]6릊C`LcM$Tq㟏G<0?sS9ߝ=oxyS٧( MuJ/ɞ0 2 pn i[5Hg1 \ǔsϴdZ}dk0A9,_ʌ eҘ  (m)k1Ĕ=L8Kֱh;ţژ2 K* %Ub"\Xau@2atӿw MbLKk}?H,Ueqc2Ȕx7sO FJ(Ֆq',bt}r&k+=]\_ܣQ;@}Dw,jdsʠ?Դ^uC?{[j4/rcEٕso<ns/=Q`4$FKJx Of<gӿx~Q?郙irԼAT *dBUV*vl0#Wa_Սe!z*N +>=/_GBȓ\樌4 H譲(c&@bO :^d؝wA9M$s#BBʒ ϑw9ڳU%1_ernϥrWKE։vҍgfaH:T`6-E9_^})q:߮պ˥+ yhkc I6Ƌ>x E] b̽,/||UA1Y]]ϱbsJҨt/sw77W?,+r#d,N;\ χ {ŵ.%,.)-X~ѸCvuwȽyl;9 &p:pP uEzu(sW@zs>8 =x}۷b4e>W` &kD'53f œNN8?*k睜gˑOJtz)v@)(#6EO6"Z sEg-n;˻㬗zw l/MGۏjsz?%%J.2'%ޏs jt]M.A_nsdy-.jBաEUf+Z*NnRWs)%#= ݠ!*v =7k+\j9*榔ˠ'NDnK<:&)X^eF4 : .|=U;1FQ|[ ijYdTltQvd|YZUUZ"=;'K%SXU $ߢ/s){-U}y΃ԬػȀ&[jWeEFE-i29UbEK`$ F^N[jiY`8| լn`M%KRV$C0# x4+4ϟ?7_$}cXм\2F|Y1jghp~G J: "#v/7o/@h,Nfm^UkzJlmy콶9DO_>^_Uo~}؋/>+<|&! ^xw_=9x { ()HB5{!|br=#,xNO8{{;c1C ]`KVU 7o\es uJw2/7y2Oc~\2Lҳ%2`a>{s`>h*GZ|}oM"R)ͣC?; u_<̽i7>/c 5dXTțGb~E7J9 wg<]6z.7R|+47{l4+nÁϥODgEc3L9[t03-F`ḉDHώxIE$ ,nAcqP[Ս͍s".Lol6XTPgܬ E`_3j*Hm6YS2n/= 8'="Ä]g~l x2 '50jb#0?^,G-3zERY)$ ZNܑք ٭)#W&M蹱6IK3BP[&J%5h`"e^ʭSNqԃ'no)hʨRaD +Qb}8'tuB<$¬~hw!zER :vVwv%3Y;iuxPj8fq?%4HD`ux[,ۺ`e֒s֩yS$eVRŒ@惌,ooG5XG_Y rP s^r6X0dlVpo`74P{8hp ;l#̊0xP(&'Ԥ#\)q4d KH PCh/jN}:RqvXrί< 㧫۟m, k.ߪ{Zu..?Nݽ[w5qJUOOv{>`|2{m y"Ks89h<ݞ1P Nz/V !!\D!I1Q/~EEB*1UUt/Si,GٴP?:o|5.T/)$7Guiԛ媋kSHd BR(qɚw흧m]pL ;O;KJ13Ca~. $Br!l]ұhNjmH/$ﻁ泺@6ꐞ{o=d«<f;yw5] }ncv՟-NJ&=v]h'&sF@6p!\t;9\/6F5˽'>/ Dzّ*ȼo_7imvw'pG0q WCf[˗4 ٪dW/jq:JSHaGVs^K٠J  Ӿ{8&kq\pBM9]D&gp2O$qi ]61\ .L#%#O69y-`b'6T"Z Pa^M%P,{ev jf,-G2L4Q(yB!`j'ND5TcT.i)B@A1"3A"3B#Jo :HMFJsĩTNsQl3@oIYaΉ>bIO!799 sNhҀ}?Tdd\AЪ^~z (ۜ]7Btǖygui]Hc2G61]42KXIHoOZ#[H2iUE:#f]T *U ֘tKQ8'{Pv2#/)?XQ6JAX ) L=WR][j|@; uvl@U_DFD^_pHyB1z &bt#BebTV(3&O)9o3O;w_烙c,v]}|;OZmLgW'[y_bZ(9P@g7ه3?f߽X&רػO^>eNpifl95NC$ fʎyګnN ÏAFG]ywqz4!ć0nnlia dTC@h#{ql뛗W;Z| Ӭڂ2[02Q3q]?SauP֕\*ܴN2q=AN*}r{]rpNP% p(:Y,v|c'呒Ϭ/[Zc[CN yFstiPxXR]u<r' ĕ2hMqWJ|d/j1<{!(3"|&,q@ZTabmSсBa2W[? ËX]!+qaQ:o-BH, 9fԉ]ͯB`tG]\"^!}E&bqcPP]ܦO8㠼Iq9NuVRNjA |/ BsseͫQcဳ,!#x"!,.m.n5ZQ}RJ;Ek;mG8# 9!Kp̧/ p,Q"wX= $`NTe'8sY``cT[vcaGR-S$?C>R|9n}XVBš*˙2Z~8_:^bGUpq1mTȄxM%Ɩ}F^_SfNP؋$ !`ОΡ8+e~K,V](y[]z{I<\Gۙos|̠kN^\zakGn+…RhUlzTMm{fX;B%y; GfpzbI*_E"> Gcof-euhMs,݊;G3ßy%㝶\nt+ 0zX0jo%? A Cf>ܵ;NtwufxB V>"muvR3%;dqQ=;bZ`y?[E_,UȝI]ń4='J,Sz »bPҋ7 ѦiSGk޽)lއqSIkTm$0uŠu~ItѦk`CGg []O1qFtYn]1(qh8hѭևq])YSca{\꼯cpv F l͆N6()\| 2'E'sR >ePީx4Ǵ[K}XVɚ$a%¯Pd,ۚ664JsD~k$Mҧ vsCkc#ߝO]%ɛU^=( df~ ~51erB^f(H5MN J74:@ıʹ 3am]Pc8ޮ 0۷B K_hgs "tQ 8($˜M 㳑DIc|Ir!wQHX4~{V?9d<~ڼ!2KifCFh(#I4Ĉ W ,QoiI鍮¨L0 Հ=n{)& ڼ}(}(BW:i$<ȞhXAQz*b ףq 2RaOiaL XOۢlB97k6ڋ\y{$;rdRckBQ(v9J<&>GQzD>fe%X#TԄj$6|~Gv1zkfQF#Ƴ |~ ;ߨUs8*eРH}` 0ۨ<ܵQ mx'41_ÖK{0D$,v'CC^,!1 ; R Y]`KAX|jZQ%>]lH-(hbPºMaMx0+Bb><䍻hRpd#iʎ#uĠu:~Qt;\u6Ut(ZitC޸v)G_qnF7YR7l%C "%95%k 2e'RHB6WQr~5 -J`~5??ƛXt/_-vVrO" EĠHVa P59k]L}8K?<g2__`{l9.N]wZ/NIz·s𕄔blꬾK[$FEΦy/g/Eޯ-^|۹5O9_Rޯw[?L|%,?펜=bk8ߪ= f% GI~ACt[X{ej'.rrW(`& (!QgE"Po 2G:`ϝk2KN]BIr/6f06+RF&Z'`zRV ;/K J09ђk2H0Y"^K 0Rǔ4^h0鸒59Ia" BO1DHR)Z9`W6Z_rNokz3LdV(6jKh2D< I'hR1jaUznnf~k͏L]=>[2bn*Bb3k/[k iX[Z/hw(?O_/j7nvh }C~yݏ`Fge^zԸ> K|*o+C* 3 w{:T}?>#>N]_NvZ^+ȖWun4-pz] R@5yѭ]6c)~ lO{iBr~ft[fd2ڦA_G4}#C"jm8à8ICIE9ʀ(BLGa(7``ϭFsLE9DTYM']]e;][]4V0JzA;<`uucsxR$o+ؐ_Ԯls]YDhֱJl17͌" ?QګiIьmv֜㤹 [LRփIn.ipCGv sc׵nYCw.jĆ7kVuPK)^_3HW/p&t)K2EAos[ Ͽɢ_ˋ#Փ2~Neg*^E^M)i#g%>]d|X}4ç:m!81ρ1l8]饻iot|^|ԮdM]T¨1{Ww_б[MT"iTeCPU0. mڶ["zap-(_/2oQKB؈n_d-ޖ ΏjdaLD eSzuDf͘d7ʘ0lmbl{n~ "%~ڐ FI| Ws~$VT(Jks:3 log\nN !VqBelbX|Ry~ܒ7 mdĒHH2J4/yCkt+ *W4%ʬ /yCSz熎8[T8H Kb.E,>KްKyV<,/'}׾8J BY-VJo. wQ Cfz1kgv.q(h_HIF!Z]C)##Q918*}w w_7Rp BȼNښr}R_KrVw{RԲ$?uy~/֥Vӓh/K)D%+nnf;~_jo& t%ʀ2_7=xF*b$;̫ [3/lq!9Tu6 U"dׁ~\EALb2VۊEKr2 yuJްu*OQG3I}!'痝3/RƳumY^xΥ֢F|饲K%KZ_;Ti~H95LܿH }; Y̩TVF0[ 83'/3ɸ ;1+2&U!#D8GB4}΀QFH%%Z)'6%| L‰ %+`)%98NC!-B%%#@+FzE@W@HAhAZ cb&L8ꮖЪ%)z261B8euBTTwL$A\ jcDʂО7")dYD`vG @^z1';% 6S“r=4ȫiX%O05jvv.uM[l-,[CmTt&yHJ6F/#PAfAYd,fiu,#-`1Ȋ-^FXGC5ǎkʗ+>1He8~Ma#q~9Y]==nȐ]zp dLw^H7 ջ)`A M˴̵2Aw7 vl>W^gcdjFs ]4\ؘzFH_~*,]3`-hKwH W`*e#Z7ף.Z^5˓~rVw(Zζ)#R0f2I?˿YcWgNi[ӭ|!>0Ѱ5ݍ,Р$: !DD.61q[FD|4GVb7GKɬL*AVT4PseI|ܓ Q#bJ` V_æ"H&*8[sKLQM9/^ -gc d%0lM2&@4T^ZfòP"fVQ[KZHBmKB,Q[kIgЍ2u+ZqA[(,ˡF|y^1TKPtgns}BTJV JI)6vQ[Fx0Dzj* ^߬fҵit}Z7먍⨥ZKAi)E!b-BKCf5>j3Rj,ԦcO_)cא~둍$t\1Έ{flAvCht^ AY4YPj?}Y9e8g^rwzNuvy{m7_rӤhWiwWyO`sv7 sOߴhKݴhzcF9$aJ`1o'q6cֳc89SH$:Ov%]U{W7lL'q( f4+carc,vI D۸Pr?kVPbo!)EHf;hO~@r Ű߲y%Oa4 9An>S9c7Pj#6LR^"Ra^+T!ewm=r{)HyJ @sE.ޝ_i2cn]=n?rOG֭Ԇ *[׎R?5C1>XIU<$*"6ɣUd-cqZ4d0y|_T̩IՎOߗB[ZoY;OqXq]Aocfgs/e pθcfԀFUyN7Rj]j\c,Ae֚I*\dm#G}OV(3!}0EpM9@G^{9;g-^e HE clo.^ rJ84\>bt9͟='{NùJA>n9ئϕPPׄL尜?W.Z#HMᆕ nrӬa8˰ܵLK=O~PdX՜>7ZQꩽ''bϽKm>|X2㻏sO xsajlc֑S{ѬPptAstJ0܄ܭ2yd*YL%#ͽ Z b^V J 1uT2dMP*q VgcCHs>{6?qu5U XQ~&ox\1 ۬kF83=t ԏAE/Ď- uz[? $FKcL`,Q<хyY\jL63ͅ'gDUbFxR"IHr7$cXΡ8Euhe4X֢lNO?6-(TM>d(QR3#&[Gyʪux>قJя\Q.@ɝysKh7Zv=TWѺ\`t7:V]~o̍~t#B4JMFR9ګҫRc914Z5gjwG%ʄu,Ms>dM]ijj G ACCqz}\6m6)`vyߣh{)?QA:{H~!\h+>.?Oܥl ms) jhCyA5ysX1}>:u ˱lV=%1SeĘ\(z+v!ؙ1mM%1kBw{9|U b*}ZIrguV Wu+ BG均;.5)|6{p9}?$XYf]_8(a%Lۊ}_bߖyD4e!*|r˅0ҙmpF:bwA;  BZQWo4~I4~~WH|KDG,pzEZ'r6[2y=NQ]ϼOt4JP.wƹ8|L7͉+u]%eRLy8|;ש@Nt:L9y}<71I5!b.ˆ@mmJUٍ0s3(lfL Ry7LjCXbϽ{EI).^E/NaXL!@j譙4u+\SW=&uE`QS ]۵Ak}t*uȶef]vԎ<@lsЈI"seI]#ԉ-fVMmy2Ե%fϨClj.Xo'ޚH|㵋tq*L46j=fynky[<'sR9ZL7~}޸ :TFW0J͉Ԭji=J R#`\rR F鵏R9J }0k?~BoD?}+[Iǻȟц>.~6~yGOxћpH1u!:.6'̕sK织 |燏!_muӮ,RU|~_˿߽k?lS4?mѫ㘌/k(d7/.j43, nCl_c!erz/:irLgV!a*87Tm*x;hѷ&Kϟrprʧ ɭG)5oXbvxx+Gfz+]WZ'jQ֤]јZ7MWfjHolcn2խ;ϓT˺>LWjV2.kLE!1s[Y?A&\j}C3j)˳/5n!+.%7]mϺ\ oG\F̮޷{z_(, z8zNI!{6v1So9hWBp Po9䦬WUݲJZh_:,lVg2^™x R~hF)EyÉ*=ؼT DXU}`_a\oS(tؑ%V3_]ߣ@C͛#tjZ/J, ŕM‚eb;e;lK¦lxI7%tȵ3aCUQ=ͮr'-ڙe^3gk+ ZE_S,. ^(5170%f}iԮrmR:mPÛG1'&p@ nl@xv@3h`Ѳ|<w=!AY_5x_Oo#ׯ5.-M]u?G1JubJt݄Mzxk3mrX&IF5%"ؐҡm3koZƐ:_W>X'wiFL+d]$/!/*5UD6u 5U kmmL`Tm JUՌѬ2 {@:0}@YߒXʭP.Pm98q.eLU?2Wl?9yWg5BnB%WTPhs/(9Ē9$ݮ߾"b.ݦBVO 8o5]'%ޛVo*ERwVXMUZmqɀ`4kHXj|w,5e(Xz cX0YQ5Kܓ,m*7 tnQLڜ90CM]?7\oɝ3+LZeJQs؂>oߧg=sى1ϥQlt!&DL>">8tO(BKeMhJZ͹YI::j.:de,2.uJNfTѬڍQY>3Mkl+~K?~sr9[z9OP[ u;w1Y֦hڴnFol::eb;1md7/}\&oe6%rA̹nm& DMWDmM6Y @q{!fk-&d/@3^8jd_ ,ƏCʫVl=Έ- К#s[''Gõ0gnử@񗟙as['F (㋉)!!%C<]x9c7AUPY AC[ +$@ ]jsrEh6}q*uN M>q'lT+UbP:k[aEwdr?+qFϤ&^B^淡(3-l@\zI"֩;$6%8ɦ$l? g'䢉"z~hd5V]hWB9 bIy.>:\NH%zbX1vVsc9mQ! 7,JdQ0jdum X7/1(HFXvDX)9I8B#6(ۮkF܈NR&ZZ?Dv:aR-|> ٠lfvD.XSdrK"9rOYrgKB\ۏs9E淯G,(Ə+jLGJm1}(vv!bmCo iuۊYm!YVgRWI`G ȡjbe+3[`pW98' &Ծtm5Cuh*][#V.hTI,1jP߉ Zkh j!Lu%vU]iC&km¥WvL{5okG+D{ KW*ɴl\YGnJEzyCx)Ir7ijRއ$ViWR/pXmʹ[.}SwOk}vmù);1&Iy/Cز,]'E/۔Ie89Aki&jrXKM(BW RާpJrKnIy/wfNkSh4&?0qR?Α>Z1Yr SoV(]nSy҇9 @?/cE20Ϛ.l X"*#׭SW+A0]9\s)- ?3w*(AYE~~~{Rz P*8qI펴п{qI#_{vҸ],n1=3f xױ}~\ˎdS(ѱ#)djjT|Q0dB+V]'䀀Zpz@^t c_$z(x*IĐcTP9`Ȕae Q.4:uCQ5j>k{fLC5qmM$WpErEvnb/ѩ&5J۰\mҋAݭS(PX~y>΍2,Vka3^[UAU}UY*PڂT|@VEUWUJ.pXZr8d̓cĦ[18amI&쏾%VPO\-SY,,VrRB7JWuE3K)Ϧ:{Lx?u oZ?b wiIk'.%U@R'EVQ$}B, gM7[ 0 \ ׶M"i#Z9Mد;:!.Mc|ǘ+NB$@t7D_%Owmi( $`Ete:V>Cjv'G2gn{J~޲"DMIJ #(Ll.;a qXY>LaՒ)*!­ho7(?W(Z޺ o^q_H 9)H.]u"s?{vğ h7w$!D7s"\]ѫH 0.aHK x0vo+v:H9~!BN?[x!#lG?hi<q5&?Wji\%*8ny?Nc42Y?P`󮿎_Z?F'Nj->疂RW, PHb%':'i}s"pH[Wl})jXܷZg8s,W0W3W%Z&u`Aj;_|B::S{5RܞuMo`ƼtW[g\.0bD4dO*9=։=!K#p׹CZٗᇝ KŘZZR Jt+!Qƕth}mz6mt"SiS´.suH=-LզlߕEϦ,5jd8Q_Փ7G%r08r1HQ %+N;NlW8 #+JmU^"}b0\I WhmѾ^$~ 6.+}A'uܽ.RbC{o003EYJ;H[jmi_69,ҿ-iO)P\{9g.(̸`q8ӧsi͌Aq2SB>"|9FTP%I=^WJۦd<}ܫUޅ_.HcC~.4ATPѷgI}e"~2i%嗯Jr9~A@ bJJrMEϹIq%M1iKѸK-|mҷm0,6p/Y'T{8|Ck[ۣ )$҇e@3.|Z\>,>i-ߣRk@5uj(/S3vfWm Ԧ{f.UEl.O Hn 8w'] Uq+^mV,u.y"sX dPFHɰݕZڴ] @S)6H!Vh3i1?y91Z'"Z%}4[&O~Lruw+Iy]|Yfp zS31H=~L! ꉜ~7|֯ !HY=cC0qIxC a{~ˆ@\8OwU޳o/<*8#c^-[R$ɢmf?7Ƙ0ǧq>G|%~ҧ(׷;Yek^SR fT\ƴi?P`ytQjjO<_}ŽLeQ6Yp+D RƺG{w.g9VǾq! 'G'UôyՓ-Nj*&/ܬ(h㺉BҳކGQ `dBȖ{d H4[Ə0zkT (vo, B[h WЖ2Ml[aRmtf+5N#!HbIH$n1 CQJqe (rx`#-1#)a{>$WZlMˎlOhnKZqx%;Sq,Q,(N!# 4QBLi,5*Jf.dn,U}g(t;1I)`* H 8Tfa3L EqgigvǂYbF̳8tx8OqظqKPbNdfYJ#HAH$xA$I#JbbN͋fH4qPhy:ZFЛ?BF(ﴠP (N93x&rG&<r|M.QKs\:W\{Ha#BଡSo*t` ?{d1Ǒ`K\b k OK^'1tK]o 7=JRC5gBr3BѬbCa(uj.8糗hͩ6wYP\vܑN蹇X "V =J d]93Sfk5;tQx76׆a8ms5U`"**&f\ {[VTWq$sڟ_xtds>y$d <%}0,زVDdmU뽪Ve  RójUw.^r.S6\ivqiUs)2B rqWKHI<%`RH(Qˎ6za0NA:t0rFmҷ(B)RBL$Ƃ@,)E&`H#ff_țZ2H11) %xidŠ) .p$2Qg)8 [K e)sH"UTV4}H ;\3ct1$%xRb$2 !)~eݱߦ>2c 1Uerx_*LPy<K#"  >׬7@|՝VI|>7_+jc',#L ),2sSZs6_|RԄzN t3 7VYgoz%r|5LktCQx˗b1I+,@KxvBbA,yky; ;qɥ&2wmij%O.y%u؂zu[<:b4 BKF"p\/+DM$n<'M8H2&G>Eh?~q~,@S6۞?8Y[)4mWrt^tV+96Z.ǻ$W`TkX(Ltw?>4;\2ԝݐ/ |粍K׉NQk 9CA>_VBhZNrPoB>Ki! vzrs}͏*695v5bp'S+1gRLznr{_7g LѨ}}5s$>[tGZ τ{]|x Zk9yاMhϾԣZܜ5!7N:%As_usFu\ۨc:4f[Z&4UI$|kd!W#Vc|WD4ѻQk߭ q=)j䈽WK_vu8Qў|0[̟͎9 5VOw/~2Mߛmǝ#&(|-KYY(-e -ˊüVxz@+| _uwL8䕍]4zFr$<깐=<{_pmNXO "G"ލALH,JfIrE3&*BQĕ (_lB_PУLmɁv&sf" g&H@Dir3pP@-)qH@!qKLEMWqɐ T F s1Qo.'fHcCʈDh0&8M2 8Js@U;e;oi"@ i SRHDшJ*%YL1fD!kmF/4dm=H.N\v7{F8Kv6wȖdnl˶<@G]_XU,Vz"oEXTg͝WZ:ca=ك.J`m;_. q]7?&&w(h@ϖ5?Zy2U_/Jw|u37kTg~Rsn]Tu0.dTjܽQׂk݃r&fA&ḽ K`\h6f"o|!,P|%۾o$?Ɗ0%&tQ qݪ8~O wx9#6BޮPEh,XuFW@mpc*<9F[˥`"GĴ-ng7<};MȓPW;wi&}iztVҥD)Vt9ua\>c }y}S X1IJ-O}w"ot (Z _O4=$J4"{wv'E G +B m9'z؈BJj.$oMi+) ^w TQ];z\|kw[rw<8]hK;5V ){5y0[yѽZ}=%+/S>geԝ^.40ad`($+qrev;uO?+lUQmtp[)|ŭ)Cf 1#$ͧPIАJFB`HPZ ufd~+SL{\4MHTxmbS0GKYN>Pd''ZD1>W+c6է_G~+Mq;<.O`ykĕO?!@ 筱O>H.HM8eʳ*dEw*=J>'|+|.TLh Iq H.PMq#H|$Utp2;rL$&~E';Ӿ#>RL%Ta #@h OL*e\L [W~wg?}~^g)i9~M>$-Wt~/V}%F%bĪOO0$.& \mSԌnDy4a%'\͆ﯯXnbIbQ+Ab/&?aoŠ!+Fv{.W{n,BLjϭK0yzu_+Goh>+XsRc]-jKMN%[LBSDLʌbR[{%=vi/飞j /"p=v Ni"=a3("QOg%v9o#MhYO^k[ӎ#}"`3d FdJi. fY<ԜcH%f((2(L| shե>''`0Lp5J,fk$<\TciDC" T[(DeXb % L (5Dƞ@naݩHy43B)},Eb&rg캹_ޯ?o_ЍdWĒpmcKU_O25qg h6ЖZcg4;l &%>NCJ֙ H5&kg Sx3k[-_[' |d6l}W2LO7vP&5%\89w4*f"aLߚf !k8d~զ=5% CK@d͕wSӽXMF+MS:ٔ?;iJTf* javõkj viRqGngQr (/(s@rv7%,pB EpsM{EpDPC*L)ݴ\@4R7ܜjb_.!OÀwN:L譧EӢ y"%S%v0v䁗D#H f{YLe)۹vNn-H+(dvXb,|D'!mH/B`_?kڭ y"/S C3EG3BM?2qȌPśK7?7EN@`LaDQ723Ûљr= ]}D晧ʊr<勾}fx˻LZn+8T -a--uOaT1-v2c9FVvmý%!% =&!6y$ӽXMFKI!rJAl&!bGIqf5R"q ՝2G )xgLPVTYzե.C*_Bp3 39 O 5ZҺ=VU tOg)Pg!Y$ѽSM#'[^4{M5/%ywf$ROm $䕋hLIƱSh7 j<wh1=DL<ڭ y"%ST3#T CC+ȅx :gJH 2}I-XJGK~"6<%p)íy!Iǩ)OR12u2e]2 eYhUD*lh.{K9/7Ώgvy/\]|J1=G,7k =39COzm5COFF-zND =CV=:~/iB7PR{=L"qt)__=^,-Gnqb.*اCQܰ5eLAIF(Ѫ S^#q,1crړ5zʜvÂKԦhu6DT=.j6$B%idD2  ( i#bq(!Jb$(i!za?< ^X7eݯM*dF|ZLLj>&nS8A{*4]vdB)/t랡6='8t`FrTimH&F3+K%&L[CRxМ^ůyCkB('{[ZuxIo8< [|/jINXBZh|0zaQ pB8_zeCI!|L`:]5Zo4ecJ Ô6]oV:*>ʰ`s5}ϨT]#tpg/h|AF?Z>P$RF8>D\-fO INjhu"ỉ1]d$1mfƌZ(FTH pBbNGZ!XDaEnLO1D\pA1~FD0DRH1`X H31EER$ A@"C%$.ugT^J/YJO8RTs*I/,dVHaG/轔^G)%C9.(gg&COJ%lW+{{ X5p?݂:IǻLn0D4>kZWp=iH;y$4wbyr]Hoo*|zk:rYI얢 |pyo}]:DҳķTEFCpNro\RN^uGSz}=1IE} zV˛Yy䷇3;eۅ2Oe;Z2~5pY}d 'wٶM W5zi&eD1 Oְ CCnʴ׋(uђcaHh2Jr^,9d3vcDg,K$j|O,{=L"it%D8Mu#ۘ^c|6cKy~GZߛ1_{咠p+q!G޻JQ[>߬,=]3{1+tB'F5uW^$G.5yK&}rIF^9{Zu4A͞K)>LD= }-cdY@`c ]Z˜*8m$aoͱkT_(Obqf)H|Ԣǩ,c89v&QOKcѠO'Y}ֹC%:Z@>N4Kn[o,hFFVrc@nfFe<0ܸar<8g?" Y3厙шv24.lBl+#ަۣAO \m%( _VݢVY彠csXLfq8Njt7 ۉv?\麧=pZȺٸIdc%z21#=̍ef(R)C:ZH.ȏrz;g϶Q"%! TD $1J ~ϓ$BHi59v)y%JD&r:|] i;4hiʐiƐ._܍Ny+1ؙ !!jSG -IC2zlBQ/"O)&¹φ!TohN 8fm+kcoY +ڛ|  0l]r oâ<# ciDv<ʉ fw;^f8{NivFpb^O`IdYXI(a.S'KOVɕZq sNzu Mjơ&!o+)C ps2Jbvn36lF$Vfwoz$8Z$\F044]4PA(D0zs S [d*HB 6dT\J Hͅ'BHbc5& ff/; &II1 jYwn[)q# :zC6 B;Q:xTIWxQc"[M)EMJSEt/(I)I+ )%MJS9#}E˖RXX6:VMB^J/UJGtn-q> ,}&:DsybZHRhs?3VՑ?03eGbcLF~;!*AxQ6zy]ZO>! : q$@{d?̜mE}kj IAHfeA͚|_lW1`&uB7ƌ֐Ge'k!_nR #|68Gu{Yv^Ճ evlM/(BX1.Aǁ!PI# l.\%t  l9~>xyB:keuʂ0G(Rɘi"l XP"Ai(Ǒ`$dJBk("ne-R7VUwz,re|B`&Y'\, X0C!6bb"NB!BF J`Iq,_9 pyk֭qxxɺ w`w[O_h! ,ozA|FmAWajy6vTxN}}#gK%{,o8dB+"j<<}_rUbiS1C +it)rs38P[M̑z68ߘ)5te*<[/NG@ L.<)nF?[![[*Ng0B̸Ґ\)fqĉYf@°γsBR B\`Ӽ!2$fq;Ku?9crIlw\c Q|r}K'q$HRB80LE)e}JƐML Hw|^UB{_j8TSG=4(ŔYl\xgI|bxyy5ɣIY'@GXR.bAex@QWi60ul cQ}noz]v=[ebb>Y]06~6 FQ"P@a7:|̣ a@pVJB[J?=cW qa%% B&Y:k˝Mo;?@Tv~0Yz6׹Aj~/ye-xZXAMXJ(҄fI&3A&)%%4D2@&@e ʥyI¢x>Deu>|ͶІweLdlSK֌e^͟u"M{7He-? Pm~CGYR+abubIX2%dXj*LPEH.!8!wE]O?.QW>P(}k@ q}foHϸ\@0/A1!_[qKY<9RG\THCR=pDO}eEb:v?= 0/v3HƯ~%>N.QTap^l/O"#nn ˧LnҒaWu"쉨w =bޡ6N#T"|Ih. G<;v{7fO#eA^zN(9hFK)Z=< m ؑT -pR _&ppJye D!?xtA0"`_L'zD G ]RB`|X J*<[ң#pKbAd{eK|nv Dls#L^Ѩf,H}X\2.2 ;շh5r|D5;F`8{>HO Z iwjgbD1S2-^_E^];:lj[nXZ-SCKZn>vdSKLwM\坯ʧ{Gq7Z@ A9[S8$\8x. $0 ])_/՞};пfx>bg T1|y(OҫqaSmWơvKj-ܫf]0{MČz:uyգs۫ig-'oX8 yM{{[5Ҡ.}Y,f6LF\M'V1oS/f8g YPp] `^fg0,D"d sg0G"IB܉A]BS;^L SI8I ]B4ie$H| lss%Rlm[! HAj?fCI"-ra_ %d4b>)2$/q|'OA.egMQܒJ.S9 eB"p46ShL9ħ-~݂㵨;O8 \(Q@Bx(QAf6[v-m\Iw3`j]Sʻ? \1Kiw6.W s(j*Dw5'p͉3TэVRVh FQ==ڛp?^<;tIŗөlSՓ=l6j 68Q`C /l;؈C / #Gq>!>֙ѥpҸK;󖗐]˫#. ev*%jeᴍrةh޲ϣN@ƹEೋy!ӂ Fe8'SH e@$ `11JD_-KhB(8XQ: r <,8bQVYὄ`%DgUxH# ŧV F*UjafWŢ:yŹ&$6I%xp !f) &k\QQ ug+rwkZfƌ D G#XkX2`9"B^`beyNXiWV=f?.C"Ms_ (snj21I%R&N,&l3?(CMҵ䈠S@qN,U$B@48 jԈf)r5 j% iKW_Cc /ozg%ldQu"kAYoVJ♐T$^h$q9/uw6͆^_ ȋ.R8u*56hT@>7ӼBRIָ$b,ۂsR/9J֬?Pa$sϔn$s1Y!\1O?UQAeJ1 #rp7GbXcP^=8|n>A%8#ʏMA@O7_z:?]t<2j^^wRL.iUb46<+3dBԳJ AT8u.DzA'TfXb/,vP}U඲4#e$xȓ `107;,{ %kY7&DȖq٥ L,U0Iǎc=`gߛ#_!iO52s͚U7wb/@TwAiG1ϫt15Ф#}qOV5j>Y|j:L\$a Ku SāDf P,GZQ%4`@gTd P䉈~yW@|ZقSv;t&hLZE˒v:,vGm&Hmtx+A..FvΘdO$fVYe`F4'B!DCE$+f@5$ R $Sa-3H] wYBN)\C@Չa&ITgۻ 5k"u:DT"HRq|Gi(X٨QQk] e,6jyR*R2:fqP)a4EJeb{R`4HUʓ$Dv…_LFoOO&U<ߥ=nY|:`ͧ wW=cؠ ]-zw?2T{Ľ-_;cTDE3v_ h`d*cEd6fi8` \E㥰Xwف{A J8|}l`^F]+{4dѭyN[Nrj=Z8hay_4JYN(2UzE+##l ؒ 9PI7M,#}ǔvhoJڟ6oԙB UYGml8۩F07"@~z2saG2YnN0frwͦ0 C {7-)jM1h^s$茝 x DњqwMs~bg>@gs^& ]%9$Guz9cTbCDKJl s5QߍVAx!jߦ`wc64 W•KݾZ7%XPNh>|G2S{nq֭ U4H*Ǭz̷c\ w ؆v[)w uZА/\E'b}MT"w*.^.:Q|@%2BɁo+ d'OԞYz1boF: b{{|0xQwh/"peuhВ,psYeg̦,^pt*M8Ec y/ JAErq#(-#"ܹGC0KgIA5>&oܳb&~X_qfz1Yجkp`6Ժww{l>~)@X׷Jvu16,ߌ&~W;|XOÏdP9e]t:yllST!ӏ!VjÄAjP.I*T(Y, qqi0&VS{eX riN<҅94@.·)Pʧ^I3\XlP=y/Q149]ws<ℐC{=#udgc='qp'nt>ksټ_a;-m<,ٺ+ \mΕ+Qe4VFU;t|ЀQʏo+:=ڧJ9ޭ7_ ~o2/s*޵m$"Ajnٓ\pxGZ#)xlo5%Q3MzAKTUuuUwW^G]k{ʝ~dg5NNj*!#xsr'oxrj)%ۮqrM5 aܺ4Q/Gn@+z9;Ծ G!Qօ4kEzN/?"lPIW]dڗ} ?-!,C,$;uZ-ی"Jʖ?Y,'+z 'ε#ͺ, 0$>W祉'&IC.rF}ު4VdӾ?ߚ!~ w_mO&L}"v2]N($=-v{"gFL阅*q%bKL|.a6ڏIXSVK7ذE/tؖxfO%r?4L@("$BI opPֶ3MmX"XsA?6]80^3g6^61l:M/~5SPC78]\`%XvQb^d%>x~d͘L\_~? '͝-oOU0̒2=b/EwM809^=Ð We),EW.^.Z ì]iҘE6QH d~HC0ׂVS$0dy2g&,Wpfd1N|<>HRsY>1狄fb{N,%e|S]iPqbӪp@\F"8&xD0Єa,0ܧm㋶vn#vbzîEA ;Ta4LF`pVD`W{"iRv &{h?αD*ig,Yl7h1bb#*= U{8\кa=v[{p۞Ecj fP|lJ1&֊FIֲrOcoFɴ5>H &̧i' x \P| G9C{ڏG L/!j=It_l%Ѹum{iwtu9k熤?:HtSb/z: L,Dջ~}JSms2JU!+)!26!Wty> 'Ǡ-҆:ܣfp1>G"`g,T4ȮR=EF=vvrCcjrp<LnG嫯|,_'0z~73We-ϋ+4q7h7L)|V&3xg,"Е 6P,#!i-UxЯU#(Vr%Ř/,QpfM'̞wRRrG,eSsя.!B?s0U"KO\8 C*žv58$w0?IsI>cvZhYʛ9~`%˨mttp[A- peQqJ!P~Qj_ dA-.+:+Uoc#/(1pN HDHϹDZ$55Fhg>A #s@j%ѝcaNЭ%>c?F+";}}GQJPbB8 yXA"@M:(^Fǥ@cGoǝ1\Tdo߾mm>k? ]PRbVJVL*D@F,k.uaXQ_K_ Q 2ƏVU6su;n0pUoBU`s0aEPh3i 4kCOQ =tXgUe laxή|("H#f|֝hm+ܔIDvkWY3)SO'&2n!╏k _맠d:Ԕ+wwn/F4]aP0S?{g,˵.c$:r³lǪR I0֟'6-b0 G C1J8VيqG6( 7r"ض]P U*ȒIh##-:-jKmbIB@# ]inG͜T+@HmGuTcF5F]Rx;,۞j*qi 5ue ;R꩗Wv_(}T_RMiIݨ,x8G_T8<D`RMSj TS s]ߞe H/4}GFP><$֐ݱ:O$2ʖv攽 ىL03}R~zlX=jLK̶[\DɔG?Ok71hNwnRM5?ڐ\DT1*C0bAjC}\xj5{d%t~``PLLքgVo4cu! v-ڂ'\|Q b˳8~D~,1t1DED*`j8D?@da 1T_R˩O )lCEJm.o >oTGͲCM<-`(^Kxi9/V8h>@Sc •%am*x|K~]NJyzx|~oӰu% f"SJim:QΖ`FL)#S M\Upe3KlO̪dqme[SzUM6H? T"#7BWvyÃv R"Q5ubNEtfMhtʦS1N,o]rŴ6{nzuy8?ХW8 'q]\ w\4O㉖TyQ:J ]}mZoJI,iqoG)%]PKIjQj.oKhMāu(A胲kBV*1ZERG0# 0ܖwk(c U2 Q A@i(!P28XZ`EICdd5#[95,X~6 cSω># ۔G䱍FC bas) _RF~S1@+>#FJ-ظGc-Whl]#R~Q+GFI|!C06Kb66A@&0RW{*{@m2E&:RbZ uj>gRԐJCQT0[i Pe3֞ǻI3n2~A \<FHשTInCB `GucE!uZ}XʥÒB;cTS9UCB4;6ş耻{*I4$qJ'v5?k{-sd?|kVQ}J5:vZzkVQ# JEOPZ@U*AtG/ 2]̹R{,ɩ2ٿ R-cbgݽ'm½A:~^`&<2H͎)%SmO!!B K= Ty{*T[RߝA^i;AZdR 8ETy5,?. R=q>\mg:Enj! ՟w Aګɽ3pPC6;7քo`ͻoTGۇjbx΋zvxkmF/Q֡Ww͍hNH3]i$|i2wȖMw)$EY AQZN s(9Ly &-;Kdk,"J(~|CфOx^TǧywWL?l{7!Q-T> }GvT޼[xw!! >nAn:N;xc"Bձ Qݺ@+hkL)- Bz~jk!ALjUvF)yRF&J3D P<}i."gF)~(4al(}IꛪԊ(Wkص%P7[R(=E"w\\*JCi&6J)C)͘zԃ@)~(ͤV\WQʨJYRFPI$:Q=Rq>rϸ4ZQ|>?ab@i.KH}SZԋ,ka`?fRk'RJPJ84k۝Qz(eO4\Aq?fRk|O\<“@KRTVOHOg'*k5"5|t(%y%8JC)}i\SD)C~(e =$MUj'RPʳR Pʉd%%S- T"bG1*w'5GV(ZZ:TiQQ2k,Ii H0 \ Ɯ]eNKZ~ &ͪĪ5)j*9Lַw|0 x MbAu4/(9yosy2zv52]F{.ԘD$߮^ڪ@d҉mDcGTrLtх"9?_|.ZrVcgdX~dҾIhv_UX^nkkPd<|ϫ ? !%܎%Eј 7chӄj?V(K,EHR%NCYjA8dIle8mC9Sň`Uko:[* Y<4@8O#Q 9"K/h{ˏqol:{N1Btl*OyE m),HA杔AS`&bT0:ZJSSI#$M•%D))Pi(X=) S 0&D1RT%ID 21bk0 %P':'T):~$OMqX.7;]n30Nv!@}~Cn>~~Lr2{ě>qz҅/"bȸ??_X6x`fN631D]|80_ff?~-#1U_Mg?Oq,u>:Os'عP%vyz z^]_1Js.#poS+9^6xRO[Q6ᰎV[bcZnf>-4 (#g\;s06f7sbH'p.˺ z@DWmՒ(f3;L[*++}suy|uV9tEnU~BAum7ߩR_|=ʟi򓙭e`m. r{b/YI+ZTBS`yOnr arߏ]?1ukwΟ/~S,\8o)G?g])vjyU}w0{Zx*DLC_s_30O㩆 uJECq {Լ:SӠQݒR/\ ͏@Xw[ͷt\a1Ғ48nޅTָ`AOySb"+.w ZFpT =T>Cr)L YQ2ғd") OIc^!} <;;*=*2?)!Azsw+&pѻ:hݎwMEyMOB^9D0%qn{7-Buw4nG\EpP[M OB^9D D4].}-^X\I'3mrGr_&ʿ8yB7' -F &zfvcq_+޺}[x[WKT_!2Z*׽x Co>l m~ ̬qg0/>fwۍ#w0>^_\~24ܼҙrR_AϬDH-qlpZnNWѮnH__b|F}p{j{Y8F~.4-}'}ዿ;jpuy _´yJS*~[\=& ?e R.l^!s~y:JVۏ 8l5{2i--Xf ΀oFJH3x~nvю}?MS!@PX^Ǝ0-^AC)v9&c(돟Ӓ{pSGTufI{{@d3 &)($/65 &i߅04Գ76 4J7J`c.Ё&q^HM{(* (|V>|Oxs|k-yK^ݹPotD"t/`uҋ$Y,GT:AM 9#mI^!NX'e)a&5c ,""x;Qfʋji׎&% {EgW,73~<XN,̼~f< 7#B§rDdfS2)vg&6#9>C))pItc6t4te.8 \ž̐61g'p$hW:Qq/ch,Œm"K"SmRF&(A6BQpj1p&弦B}L2)cQ6:)u1*4ՂI:Đ".S!R&$%!RMC]Gxˌ;O | O3>O;{C\1CRN*02&3Ĕqs==jaHW:q2Bm^v\loNvQLBFμYֽú^NXZbupζ.}˭z7_/-o٭ʹ2%7)iet <&D)З|cbXu |"Ip:ĉ-G&&q$_[' :!QweQ:1w}{5q^Yj5VRaIl^xBUu ='Y=.G<:F/ 3j0V$ZV!(ג4I <\.$ׅz7cףߧ 7]x(ٓy~6.DU|^3%JQl;.@7e[HUtttt[R==P铬G8:"7 pLh!F{\u5Sc Xuu6Q! 7[# Ss x_\w좦?櫿d[]s<_ϒ=G5?UNy{7w@"TbW._A4\ frp6-xI"q9wvaW녣z YS}Bpxg S*2?;!Azw+ѻ:hݎ{9[xrw!!*& 5n@TuL/w;*+n}[ym)ޭ$|u~ ;_'S9D)akO ""hmlJ̢є)t( +TĖ޵6mdٿҗ ~?T$kWg83Z4$ęI5AjAT99}>.Sd2 UhCw1uVmҝm,»؁M9b |5%+8$, u?jtySʘqktxXXLwYf׻d/J)2gS-8gEqEeeNjuX4C"X~245d!tor;#a[^0OO.<Njȷ7GnGFaYޗUVϮZCi!kFxӺvz'W>;/SsD)tݱӢRZ!Kϑ'bnXʭ̋.YzLiQjS/c`Y(7Mt/XzLiQjE {Krc)Cٌ/T/XʐK3j`Y7r2Q {cROR YJ%rK-'vNsdS: {cRORkIҳf)n,N[>aRLjI_z,ečxxK 2Lj&z:orlI/XgLj3Ka!4u_z\iQj谫,%̍e^/lqEKϛgX􂥔4ZI=R&XʄaӢԚR=Yg|&3iv7)7U*iE Βx;_BzBiQX6it`g&0/pJ.)𛝰wfoj[C+Dyoǜ<'>~}mV!C"]CXAPD ,\wKm(y{49e5%ڨ).Al bd'EyuOAbL4s߇woV /cL<ղ0^?d1K/lqw_zW>IC*|"ԌxBM )0O78:.-S EB&;u.eS99\5z8ATpaLإavf^hA #>ia֯#Z'Pp_g7`t G`3P,&9~\ksmc nM4U Q2i)Fi4T0.4’jSB_f43.<ִu٭ѺEC@֭dVbYlfRAN0w lѕ}༭ĝڡ(Ēw#!ńp^N'E&쀦JǦY7A<ֱ$&LgDbM$.BhX([U=gn\omCN_fXOtz5lW];Z5_͂W$ nVoun0FY؎$ O$2iXLRʐ" A$$\kpX4v.8];){kӋR\): L!(Q" s BqH$hOQr(ˁyј)8%W,>) EQ!3fd1NД#zQaĞS`R2T*PHDlCyi KɺBSgJ{jԎj&<7xPEg07tI# ʡ6CHn֨KR57n`E*h&(BgDjTET IR@Q[Tʐ1/5j{FpkyG SÂT8UQH"Ct4T)N#xzBD_d}6i7+>cކY~[w}(~o7"ʹiАK}>%7/D& E}d_]` ч^_m^,WkD=dʾc&kW }1 oD+1Ja>{/Gz)uF7D*E}b5~G'f[]jD &#HN>DO^B> ~B>"dJ3Z{>20AROR<? "#sd R:Pצѭjкc Ƴvm Mγvp; '0!+ع/kj, AoWwlѣ!__G΢dtq Us=cV.],\G56'\4`o*i܄0B!f~hʉpܥ''KP=*,UY={3vz81n`}h{2RY6 e074QPsXc炼>k"nG+|q^֥Ol?<7cBgIYFB&TMuiɇmOq+WOoQ3}P{FGsLOjO 1dOMN~Hś_! )4a,gOG֚ur],<~d<_eAK)OV rWbjYDW{ A,׵bQH0L_mB,3P:`>3_n$f3]ˆ@VB)B9VH#Fq ($.okt]^к\PNJ-NVU9lM t gEfR+P<<򆾾Ԍh~)BQ3Hh:zfGERk%w./L`ȻS|wo^^DޙZs{C>ݕZZ͌9Lьtqw;Z/3a/H,`q4uVV~nbXĭ+muL9Gn֭ pzq{{ c+C>una0bNm S9Z&)ZSmڻƣD rZ[cV"rdVJj7ѱ~{fJo/? J1LN1P i'SVw'-ٞAGC(kFjlxA"'$-axBl(t J`!s"x5XbOzSb,( )VpiفjbZXMBA׋/6*TAݣI^[ȽGӱ!6eYMssT41 aF'/1Y/f|O"Qa!XQE*) FSRR++\On=]~OWQ8q@ "PO :t(4UYS4 [/eBS&R A_L ڧHôZPjH(k| Pd'p_g8 Pla^xJT*0적Dݭ/AfLt. H-*6C 9hI*L# <&!F2UQcp)VM2 (:DEUM\'vMd!*_U@)pq 5駑eSđ D\)Uę0ܙ1KbOQdb8%y: h*,TNP'B*$iBuDD!Q?q&{:t~b'te˼~ؖURYhז5!ǐG3x1RDB!:(5.: "E 0%Zɋzo><8 ;n:n$ qtDBwuZf6}b>y1:bEMt5>&tOjieg GwdA/E1MZr\1Y'µ+6QzyIw=?b(hQ0vc)'K&e&up*18lkoU2\+iVo)aGC\Df$})\pAsT-|5ꂀ>FdloU-rC;f[H SZ?~{t{BO$a1|7 % VI1& VKu|2+栫)L8)0*[vd-T&umQ I0v!=L&0IԎ7T${pJ*J#%Z/Fsa(8}*V@터j "Զ*הҞ_H nW5^imZWrJ{ђXieq_:b4RiSN R&WUVs9Ƣoڳc{Ij1.?:AcWaHhw`x5*`Y0QQQ~WU FgqlWen勻&b;5!4VRwR}@;F iNSqaM8kM/DޛiAj겿wy>RΦT^9oǵ\ y┐%/>uLaVoͺ;Z&))zq[7nZ7_PU[=?aѺ5N#1z5+rԱ[ X/±#v!ñ)rc)E"i/XJK31ACf)stNI;SK=-J*Kϑ\,9#;)\Nxn vt]PrH $N$Fi0֊s ҀQ$A8QnI!C"Bv]CRޙ*(vy@if*Wkӥ{ܟNAbO[k+Kup_~?+!Ta٘)ASbpX

I?zk&1PnS؉KKwCxf=[8薉2󒩊`8>jV$_D>9Q+4H*efb#a~j@rE~gt%S5 ]ЋŇҧ~^y4ڏv)7+Դ ;5Cةǧ j;`>8qt0n9Ys<[gQI ٩zw1t53Q@>fw%r0x3i1؞yő@u8:/@,mF/q4 S.TA`\9>kI|IL :V/CK/CrsDF@a:&g<'n9$hKu,hd᠄`3!.27(8MͰA/X7t6(u$lK-5 Tϰ`?w_Υƿ1=NKn4BiPa n襠e;ht7_ p* TtfH;y $|Qb3$4n(24Wu GuNaEڂ !_`<@%%8LV{0{š q4x:W:5K4S8:Nzzr[JiI^$v]]Z}ӊ[\mDbOzw :/o `zz?'ʼn*"|?H7xa$B?~|dh}o'1K?=L1`Վg?{3ŲR-?݅š | qxa5}uPv;GII,&U 51hxoʹ rtW"~飘BLnK"F;^_9?, A"~G^24hƣ$"m۱jݲ9<{u[hz>'n^s{=BkYՌNmCmH:[:(yO&: k<9vs*L1)Lvo>#eleU?C(lQ= 4WjQ$/pcqn;}I88! ?wMXI>q#)CB(jPptv!<;STP ӽh$f80BM20lo<6%VPm+ *Mq!?Ĺ+K_&wKxUu`'m?= &]ׇh&w:HCE-n=_owENB$ӣ?-7b?t-W%E<~[=,d\svGX#]Ӳc/7Y-+((OWh$sٜ>r20(8SD,GKCZuqj?|Y231X &xdP"(t (pNnjV\c{+TM5pO^p!^#/%5.;M\mO@!>$OwI8p/ CXL; Y^鹶\Q.^Fi|0 BY VA{ [": ;3%II:$a)^9` oWQxL "/-Rp_}$ɟoU2e.L$kD먤|V_ۙf}7S3"Ҋn SKJ0\9="|_g8wXf'y̑EV9NP\CR9>7s$"*ERb* ->FZ$y+F  F+@+#H$Qƙ*$DpJdqgS圷\2D/:ڪJe Jv!RA[!9>M8gSӪvD^ãB 8y Sj^N|&iPW KtFeT{)yD˶I Y9ʚޡL͖N,іy5o$;^&8F@0a֜ᑮ _%(7$Wߦ1VJUҏpA0%4'X vw4ň-aIPHPk䚀 Z" k΍7-%Zl{H%ȇUö͇mGd &5|ض`,nFJ $w Vjgpq_:pS vKu]1xbQV3Xs%#k[ 3Ra¬`Fp_zojmeTU6R@>v>'`Ԇ 3!w!Z^gjy!fL^ZRhVܙjIX`'޶mBɣm;+.TV&μ.dE Xu1S(?! 8=UAXpEq+=2se6bcJS^MY;n1RykETNxnJ O{%K Os('a =Ȋ.$d+4nHnM94 b8 z`4^v |Ԉ\ķȕ}4^DFh/GcC،(xƔ K -UDMbXlDy I9(Yx#~W6QyڒRض-Я tQS3D12g۶v]t\%'siw;a2\=HPRk9Phv?GaՔ#\r$=|)G:qxG[pսǀrݦBy^:]_*]vK[V%c*cAq;uCpE5Llr{ Y$/Ժ\bAzR<]ݛ`'DK *:Q=/A&: /qޥp,ΘvRnTKM&7Ɉ5(x9Y%W`XA2X묥Wϗ:ǟ $~xdUI>R:+7ssu_j5u1C>pV9. M_ '#,:ͦ~:eFk ~P<оy%y)"؂F3Fzi4a#܀%05 yKcu^ $$/is!9 c,Aea|y,fZ#]H(%uF9"jsNGfPJԄB5K;.ֿӘq;)/(VC+9S @y\PiL)"*1(>xBȚ%/#i3m_vV=2Z@$_]DYRRgOn4eW7l+s9N^F׹@$p[62T H .)9^^F&eUl;12ڱs9r^F+U2Pb0^^;9/|ztmFOy-8J䵶}e[;M$!٠y]..ӎN_ b3Lib^F1QzhOhEu68eu)N'?Rw[5&iAZg`Hc?hTЮ[:h}kg'P%_tpΙ|%5/맱b2^.HdY66\quȮ2&Y6/ƛaRG^bs@#ĔfKc(I͙M=r_enk^vXl'Su}A892f&Ah Mr0oYi<}.%_;pFп~{gr[X-k߸/;ʋX뜷3/(ogN Lv 2ϒ}U^.1dAE[jP>BR53|}IKϦ%Yr˲DK%%d~9GRD{5zTx½:i=fޫҊ̛̿)Qf>H cS{Dd-aϖ٬%y2/%SǼ]59wR^0ӫ9dJN\F=3;k?\^Y{g T;q6uGthw}ihτW=~k]̽MſB:C5dڵGRoG:~֕cICg[1Y;cJTY❰ U/: oȏ#j\4h?HVbޖ/=Xv6z\xی#&[kRPQIؔ@"ZzhxwpqΙ~@_I$'{kvgEw<ܯ_Zx'p'v1B{%db يPBZQR> '5T2JdRYw3NBEnV+6W*jT%oĠSX~1ײ5[DW`}GFK i;0f 9jXǃ$G6i=`6fzRW;|&XѓiAuVZbS6: nMO!pnb 2n)iO|TĦ1kFOӕw)+US!J 8毾C?XR.,n>ϣ ,#VK/tY*cKém7`T;|fwv-x VBG]ǚ9Nxᩛ!@< d<0_lf2 + >dۧz2Y6ROx{=(uahCL`M&k[_.#jToz6pދ]MBN FF9N*B(JʆC֗&>O@KF|h@g<wB$P%$[u ΫAG>[ء4~PzVphmLM#yAiaAQ"bKA/ro3.}rk:z-R׽瑟1 =7T֬#N?{>L{ܻVmfj +V'fأZe;K8?=kԁ7;&M^:sa8z뫥 >e_Z -g>KJ}P-\@-k"Mz^#RFRl\nZ`tFLj]T}ygAs7A"TǦ50iQdTP٥ס˃WJBx56y4*'xfzg(Ϝh)aFH*˰Is<>pיLO#A{t͖{ZL÷u\N_;ل\uaA}V S̲)D)Ecl>kR@j.f-LԌj2Ю&CU :H?+v{)!ڞ(U Xs4W%PC]R]Zr]R,z_^j.2M JP$23{2w-"lwi~KׇADVjDjh[)8׶&|N3 f]I@ʋR|Op>DN5LE,ɚLɃOهl_Qt;9 '!(7 و.q)RݧtqڧAtܘwA5Hv، MMm4|{R6THg~Jx4TZڌTj^@ޓKcMyIJl=U'RjAO*L>uHB^GYdwvobKIU{"NBiq.Pg}).񛫖rqF<-p\]aR{:]8cߵ}UǹL.}lj˽1*A`"``մ5s0I(TZհ+{{1t!$[!_DimNɎǑj'dAq_$%)`ۙJLdԪ='DF_'JEԆ#]IO@eGOC)|'g/l??`4^X84z ܥ`lC!?~)?/˪< Ϸ :nuyIqWQ009>H&'!V;./Na:MxZ-{N.cLVeylף>EIW /쯺m>.6H`.l[^ز8e)}f:|_M4"`3<w9d}p.UF]v_6i)kcJ^K$jV$0}M%.-ˬG,5M5GVB * S(H@I /vt_n/=yȕ;:J"YsSpҗMwûUԽC]; R6*b")jAbC+'HT힏aji:݌.TAapp8+ SIV?Q9)''Wl=wP`ltvJg]<+%)4ϣ1Z 4 -:ȊyV**`U:*2u4b*8[FH<(47sCZbpFyIKaX ^v'+$AAsF/ {b(SHL(le4{^2 e s~W4fFVŞr[nx@Ǥ^쎃I.H"Dݭi4 980Z( l5)$7Z4āC9,Bֆn^z}XaSxԻV$ Epd " yPџ<3gBmoGj#!IXxc@%Yem/zugs_k}- zq}wc+\Ee+OM'I=\?V} 4 ڕaT)rLɀOո%ê+mwoUKLx#k< jOVon-1P EuEb g,!>bIfˣdSi72ˡ*L)oE|÷mj=*Iv̘#ijCLel d?'B0Ga(Вs95޵4뿂r6GiK$[K엓W'~l! ^8.L?TWUߪ&KБ3HCޯnNA0a/uYrm96ʍɸKQ&IQHa}v &W]86(鉢q![u!.6'5FY3G6loT/l=ӕ"7һ"IS.(:4cXP','m @P `6C[ga1As7\cZd$CP@-RɽEz썮^'. ]R C4BO.knx cG<:B{K?+\8K9r"%R G@ՎY=. (;lL,>55.np5OqZ:xiƝON@F~;W_冬~ě^}s'l3&9ӻKmYxhF1蹊ovW 8$w[bY 'l@4v_hr:ƻxmEU#oZXo޽}HR~ɗ+#Q 𱤣B| Py O:; ntI2nh6 o(c;3/,=f` ^ ufFf]F_ΥrtYo{MXt.>'>S`Ղ{?OÈ}Bֳoj2wh_%Zղ;U[=ܹ,|䜗ON:YW,ېh0@~5}TB}ûaYrsA,,[uiI958`F;ʸ_܌;7i3<"mm/sw3'^jl=sDe|w#16{+I{9mD[n8&'Yhqݸ3r^'z ZFZvytK0Y?xL. l i"gu`u?:*7tI|}Hǻ/y~}3v7'vjTtXjƄa0bؐbyˑsA{07̳-S8d҆=,fyJYido's;iY7O7'{BFEelG@VqOVLHB" JbA3』(ct-ړW~]v]b zPƪr=e"Qͻ4ۻD^K3׶;=DJB&`w"V鞑xV$3dq~tТ@Ϧ?89A-$9>y3A}ٙ~fpNjuW{i[i[}&W̙ 웳6i\ߌMtJUF1C+& O|쀢,#Vs&';=(-wA{D-GJQ#- 08͎p[=J5Sq;Q,Bm><p;aM";S$; &Ol^P5)yNqr>F@ -z~?Na05Q sl^BƝQ9>^g[\ ڮ\[B@1dIY=*wGGϏL4w4o)b q8ҁ7XZit /xC0:bKOI7vf^ CHýgQlR05' "ܓWP0ǶRT:NvE&< n>iT'R~A(m1߹ˮ)! HsnsL`2 oR7[@fT1g}(] AQ GK 1M] σ~2){;.\?&8]X<nXhbw6 A;@X맿 Ļy2 9$(kATe/3PH؄TBjΑh 4MA\0{"V*.R4V ǵfj;? l :w 춹ń< ДyvJxLX5JCso*\r ;O, Ӳ+-?ûn'W@w]Z[uVl.<[򚣫xrwJx"g@wD^ʷʷhO^+0l޲T?Y&Eb{eu$Z$Q 'dPn^Y1Gh`6&)O hhcO&NDDi`uY[}zyb`?mp`=_ٷK$F }GEF= 8c\ ѤQ"1XRD0JI2uyVm'[OoW6RB{Z*yɧ4^8\:!u}. sĐf6TX0c1ĉӷ[v tJVTgSY =Ag"0AZ8a\c2#?FWXC\ӡuVnK+8/)wjƇKHN[rLFgokIwmPIƾ C7g6\o LJ+% P{$f$lC 1Z3~6wazxf^/ )8wI8%51&R$-5(`Jo]2%>5hB%n糠 OUXF0QQ3rC퐫]W>\N\ ?_`/d %q-2s ނ"GMho8iKDld`p{r283"z*=N#EB ?iMcʰ8 hB% ZM>Ն'bN23 Qz0Ӫ0eac U)*1' 8 >Ҁ)/|al1ӹ@#, %bG gA8V'.l D< '4mi B#5}₈βRkl/a%D>OQ׼,lb QH*vU[IЍ2*k8X ,wWp)o~p3ZC4 Zj$-a?+{ewty9jx܁$lk6l 9;EXWa[Ƴ$Wk0Z1m& a]Zoէ_P<{gϚk!FQmPõ${pSA[.]18M(ӥ$,t80ѨYHʤD G,O2ʢK4FӞ)O\r6(,^c,^\FƟFl\heZ=6^gF8:0 CxiAɝb$Z:< N+eϣ"rze$bn*1k /FO'3cv\'Dsk. )^ 5h{3e_GwLk XXXj"9^̪X*l5$?-Oej=ȳBB©i e5ݻy;ԯ87OeXфws=\PZXV3㼐4#s caфΉ*?XיO>|.|Hs>v՗ E.bf=:(R* J*-/l%U֛P&T#eGYd!tA@2*)3I ( a<+3B.TV6~MU͟Gߛrj,ܱ[>u1|jl^Z2_0ӷՕg521LdYOZڅ.ҮT DZ,\QQӜ?xNIɔ>MK0620fi{?]ۺ^I308{&T`)0 -Tsan;UX+,SSEFj7\{A3vyZQg1s1ԇt? k"Iq-(ʚt]KᑽOIXW)\~Ւj0ؑQfG:a:^mW%dӹvkgR)mT>]tUm#E:PD˧[DW|j\ &piI]<vFwXsZKy(áЋCx4`&[?||JZoڞf]\\z]wjbL:}wږLw.pgђ.|װ=}{sS:,~cƯsxO}piIB(cK )Ow.d+JTy?׹"O妢ۋ/=Sa9=noF76pCOs;;eD?Eҫ>*sttox'ɉ+}bZFm'24!g˴`;j/ehv@G_eNbW\gE%lu85,/ݏs rjmҙ&o~7R-#jxCLˡ8!ox.:.+Ay}f%jN:+( +MSKQ}YZIvF[:e-$(=WY%"S{ғR2jy -<{Y'*d -dD0*0薎!F)]$1BaKpP(1"9u*ͪ1G_LpRD߮kvcelka2E)7LHUYMAgYA^0ZIgn-2P̬'5e<l@65) F-2Vv6fʓԤZh˫0:령~rCQPt@"3+,,tFT12+v44QRZ߹@"}u5bjˈga Ձ' ݆5ч̇S-֮{u +qQ1r'6r{_-e\<Ωff1*}novyz?Eӫ&wU]"vnBŭ~e]?~)S۹;XQvU8/Huآ*e%Eiͫ0Sfrt4vz#8:F;h~tRNF+@$ <(aHxF!jWO:)4Q꓏w>s8D߉J DQ8pZ mpe5~÷s)Z-N(7s"vCްޜ_$a@NOI/QHזn(y^ ChGX()E!hˀȒ4W$%ILnΑhX[ZQ:%.HP{.aHPa;(2C39V ƩrlS \rJkz(X)fBeҒ)BIB2w*aSueX Ɯ/1C;8f:\)}[=6sOta.샩F >9ˇLE;c? m7m=ڐU_/$vEEBPGtg^SzYT/|h#7&*˾Ilz}eJwSPS"/ɾ Hs1-"F\vÙ>q_X*> |D$_.g9F#ʱݺqth`T&`c< By>mrI\}vfp8v' lɈ:M0 T/p4Sq~;\*̻6H߆褭$=#ӿH (=`ҟ-S48p!^R5FFCWN#F^95ǡgyAt YaRi E ʲuF"ʑ"xfeG?ۊzΧƮ3?o_̖V٫ "5lr3 ,0`!g:l?՗umtN;>hHcK :v ؀| uw&_J&^V{h zw㐶QFQDz&*ߕaj> z=v^a+cX<4t+"Ct(ZYK같P3ƅ7 sS<;j&LZݥnSWOw.d@jMs,Q8Nw(cVБ[Gi7H~F5-(w{)Jy_WA箫݆,mfE1gg?=d,RjhI}9}oQo󛩯n9a]4 788[ buOwWCgu'P" D#n>CZ5]s.zJ0ZSwHA8(JR" 93/nfԣȸ<:=0Do׭]My'uļ}5~ |f 0C3:~RtpFM䇏@ ͣ2V95 Yc#z4H#2VKy76T2-xȦdz]lr;4u!~"7m)hp]t@A7Thٸ2y I5?f&7߮ -.I̖5FBQ"ouw? Q58^ekeZ*emK= -LK3Հh-UN2&' RA]JJJ%Ra9DRMwhCg64uID!X9ky0jE QyƯ_ ]cTY2(qRe,=da)\Z gWuXa3$C~8AwǚS.9oԛDЈ3-4fkN,:âqhYљ̋^H>(O?7QJJg7fR Y8jx̂.¼L-m߿{%1r崙^CbOo+:˻iޡ$4D9Qr8À*XGXP vlyCmRl+yM]`ldH='R%(q]_.@ f[$~}S$#·hsB&.UN׵pHL0D宽ݫ*0j`ɿZ(Hàjz Ҟ`@n Yr($&TZn e Ktj0XSC皪9ɁǨOR'd2:C&Ap *An"z\$]u { >"lvyA"8K"itH*Y$Ie"dZ[3UNf]IQ Ղo¤,Rc=[JFVh CB+ |0 )B) 5LH~,֛`"^(B>3eye;~_@RJ%0zE?peִh|U>RRܺE;% UP=Z!4+W^:%=MּdrDW%]1f)nChW{T+vd0wWX]ˋ:i$ -Rz"z8xEk,㵺&MX/;Y/zx[{WؾZ\F]55ي\ӣNs>JeXD֜[[b(09@Hx/WiGvrٽ>X _O =*Qݠ SjeoLQW:˩mw B5TlF|eF|%#63թseQ(8E 2퓹;ohe>TVzIm^ 7z> YwjdZ <_]3(>Llb) ,mͺb>&(YÔ2ؑMX^`5ٛMft~"vEΐ9gOol]]hM9NQ"Fӧñ 8^3ϯnMy\# '*Dn"[񰇹 ˒K tz H; RE@\x ,k8N H%^w"Z ׹ $+0Tk\2 9^d@n+6G]ya:)! ]M*Ò:q](i Cj냶P\kヤ)j Z9 wNNBJ DSc>0zK bȍ }uzoycsqe|[rx'Z}myt F .'&K*Tg9iUdhɖwC?VG%RmPg󘼆(g29grr0\#qńro _f/3ȗ&$csVTcjY:k1$e%yqsV}2TR/˼<y=:~-%[2몼!S2@ki2R҂ֆQ'AXD*S /5NWVQ2j6w:("쫃}DwOkCm;Z&-;ZiVFCܔo>)>n;Z!ǖAn^Y-Hsc5mxK}ͧ՜gK"0١2wq,Iln\ ڡHmwhJ80䐴VF~,y "?~ Xu=d5pA[./3hޖA[Gw͢U!̵9?2.hZ&=鲕hd-t0aw'sAa\`MߣUCN]LK&&wosn F:d j,#X/i'lyaC |sZ@V-QǶ ~$9 _ёN4{kƫ@PPVRJ_e)X|A _f/ HgdCrQ;\R@] he\R-|4z_)/[R ^0}۶rByIuȵGc 8ډAby}7mM >_:-?{ȏ1A_6K=DLݽouJ iڪgu52P53R.뺻}j(1Tfe)ECFrckO6աN!:!=)vޅu\zS|f條nt@܌p[4ډY~QLaõ^T+s\a)u1|̀lf@2^Ga0hBw%N؎-oѺdzEo57]3n&a8UnW@p;[}:E\cYCLq>B6ܓx~Kߛŧ9,̓6gCȞI51d$pN$0ugZG^S}A$ȼU,d!Ƃ+VUPXC'?7 nmCWYBTVWŊTÁ>ٻKWݻk\k'IKr; /= ZHFU DK/x,цDsܲ*#7<2z_ݐ-7wV 't~P1 FSGgq]ؗ57plfd23d 2np0 ]d&;%a8@#Z:ڜ7ztپ8$FS!yՏ7%XG4ePU|{}y[zA]or3݁E!&tp{4ORJ C (|m^ٻ6r*SsU@تy7⯭Db{ᅠYi=ԙ,۰"_tEVq~uDE:"#-`//o#Ct.LRO4@Ќc[|EzsR`v|p|= Ȩuq ,{I?_:JƊ0hӚvk}(sZcdﭶUX"#+OL- ّ.sM*B%e^i- ?\zl p| )bJB$P%ԒFQy:dנ]*uHҕTRμn\~89}LۆƁ̶=8-(JU@`JʻDA.^.^yRqy'/ewi)>,Rɱqg4 'omF27k%G| d]- Hi0+JQr#(7l*d;ʹsWJs,45qڊh|H㠭F % !(:`!ʤ,a. UgF+Ke;.ӆP5Q`[B ^D[(&|땳yp_!GQix߱&5r!Ɍ%_^Ѿ ۣJ]ୖa9n'-^8c]x_6ɡ5Ĕ'vT d]nۊe 9l$̀X Ƞ/G"DNgXNj5+lqb:e*aqJ _D9L3ϬsuJ`S i'tNSϵ[^ ?=D~9Ӝ00\B@*iӨNKJl \YK|j!y|Ԝ.a41(A1*ota\KRzg.CetD _̆zk2#niK/V><ҧ*QV-n˧Vd-zC 뵾n~uB:+޻שRSIAZDp$o}>dŦ|5 '@$C{D0 iQ*a* z)pbi+3# 1*-0@NTd:H.Hs/ f2v&.Y/e9iPvޒKۍZ*W>C <tQ'xdsdMV+&B޲ F^dfP^L\ ߉:I.PG="$ж :F\ iqڝfbխjsTz2^ԀͰjEuUɲ6~׹2zW5m67qWu\0H` ^GI*}tj.zNٱ77:UP82QJ bxr@:Srj]0Ц%< " yDWJRd ˶m ⩳+fV MAVJTGt]9q)۲ ݗۺ_L%Tb)rN9CosC]s_Z _WcW+v~Ek@sn+ZCk|yh )@C"b(O1ovaFhIY^gٟ9M,@d]ی#i˯f!gO/W׮s>{6~XHfF3F"Ť6Xf9~bu+)3g 7%lk*PI<}SclŪ)6:&MEQgsߛߎ!!يY`rr`P yWH(ȶBT]Ott6\1{FIGv'k]Yh ' }0lVyZvGzrJ$xܠk gIVd6g~ =pef/ϬLHY? zq}s2\EekMhy{_L`{xru|[1ooAU])UziM"ǔɕc?箔/φ{٣19\٫߬6yP $DIh=@`UR/B|F(%H9!9,*`OGoWS#+UGBUJȱk1n=ȨKPCyD٘cݣnv±:IКs:#M*+5 (g& \)=:o~05(+L&BZ^؍Կ&@fOv!{ŎV=1tDkOFbm7@gc;U?"'];_ W 3Rp8ދĎIO+^ ϶sa~_d^ෙ4@1e-ZԺChljٔ>.S8ywmmK1_TM6]qu53Q$Ël RDS2(eɮIӗtc=$\|}%˽'$>  twjgDko\e+P )=Nr Bۨf!Q`|L[4K;T>_C* Z|DžāpSc|I,<Ïg=.JpD$D-3"3ɳ~:@^ ":4׹`^(mHAPATNmՖԇcJl$T=HBʅY;籠)ZjY'b"B*!Qe8=뒋k.I 0*K qG-#\X)BwJ<0!6ߒֈψ?]B@Hi8!n=.--EM@\)/mifBf~zf^Obss.Y̋$}ՂᨠQv%9?2^C(:|bHՖh%&yhuMK?ƞȭ e;mb>>Wٙ{`$.3x' E YEex%И%έ*-]@k4}>F;ZIß2TJǮtЧJ"1rDtčy9ϕHOHqcpc~3stk*St@1c}B0/r̿Զײ[՚~TRB4[Bps'E#vTOw1קD ϕ#[yJbzE(tNHj*hSDeKkKp3wR~z )^\s˜9I^_skKB{X%ڬĸ`тg#\P3p`k+**$=,bSa1?5eH>Z !ub !51zAD<Xw.l.t T8j3<:a1$fl<u~Cl6PC98n0;9̊d0nQֈ0[}C/'1ͧX,v%5ȣQ#"Wv_, k{uNQh=J ,;llyt>N,9yܛZM3ݚ1%\dlPlZrMP oR:|8L^a/[Qh DW:r#M~׋g.Wʨ8ל ` wo>ʨKp=F >\>VH[\!*b ٹq:g6:O?wdZVR:!(Y ,3 )pF|M\ֳy` Krc1A u "]} iBy-uH}/' د߾gbz,d/s\ fGx?(b+Cq(aG,!N|Xt_Bѕ(s0~FøA Ǡp8m=߿=k\w<2.OS sxy׫`?fIch0r6҇N2 kոt?wpydnƓ;Łr",Jr˝K1 PѺߗRs+ntl E+Ivή`tz@5Fŋs^'ExYP2s*~ lelLE7#޼CQ#~諺<0:.YW.+ 4B~M$b2Z Yv2RV _lu#Hk.(H51.1Az9l&TrS)QdiXo 4ܧOjʼ%{"% V02mKے2x\ E-7PʡDw˝]+2Ffr@" _$2R9*(ERTjɪ7!}b췘&(S/跼004`hfǣp'eֺjvSu )cVP(Vư$G.caI Nh*Uq!A9K0 nd]@co׽ۑ@o@.F& MZ˾%A/R(sX܁[h  BD$ (<Ҹ.^qo {xJ}SHE.Ί"/aᩒ.~~sj83B',O5|qz$?޿uQro/RLgK7=V OIFy0n-vήX$Z)PI6 `uђֿ͇Ch qmjĂ_20}ē/o7F%@^!k,1;ze]Io%5tλ},uc4[aT`Zoz΃D+ : #\`Ia"?^4?LӅ+'?h ָ2#ڶ<ټGvCFؽn _e ,%g4A")W쨴W/vϛTIKD{IZ+vFGZ?2k!+F5C vnr<8ܼ֒|%gqxq \%ekZ]}7i8\w`~.߂*[c pgnZME#noqGWTiB~5u+/{p.wGUZm)8FE\l|KY^FEi&^t/A;>t>4#$1& %ȣ[R]f}մqmi[\Mk{hKV>aa;*zՌƬokNwʦ>N% 5[kAle^[EZodt-g3{s]WM?`(u4L|?C[[Y|ߔPnmlW^hE!6^ⴼ&{:<]K?Y@+J+8y|mgsp#obڜ뭹^_= +<-I32E {oInKڭ) Nwl;@`ڭ~ڭ y"Da{q~4ݚҠ4}61y0HMֆFY7fv֨j?XzQrN w N}gދ wdXeް,kJ]B9z$i0V)Mio :@2,8Aly"qQPץTښ8؞-i덭Em%zzcKՑR-)'6YyRjJjDeg]ɲ(s0 .-Y%`]xdu92bJpMO1rcbDtyA'6H ykmI iqE{>I>+ZgZ不З ̨(>{w9Q|9C&JGl9~d.ffwzXdUJg'II\Y!Awuud7T%RGY~5unKRz;sJXt4E"gI5{PZsox/D*("&A-8-^9-{_)*#;4*C?+T2g4CDMPe%1%Z }w j菤t7 ޹bPhMuC #G3L/NQb7Ru;=vj)rP &8V6kؓvq5+uf6npjҊ^ZiTzFs2jk3ikh%atNcxttCAC/-]|ds0M҅Iծta; B?Qu4!h>Elrf\ElE5on-#KˁIRbvmEfNX+o)MQۖG+Mf44 y&eS Un$.л trƻ-K؉ޭ y&ڰ)hSagf,ѻ3SVi|] |VVZ8Yq[)&Z)%!V}>!ePX-Elj"ei]崷,&>X.I5&RaPHgQ}ѤZ(sZ[Ƿsa7gm믟>glrIM8ln?{,?+u(0M4F]<1q>y\BSghDxVѼ{}\MG{շ߽qssMq"vDc*T҄q(VS Ks S,PQ#pRR)]VST5G2ŠQe-SK,|_s:[!?nQH\h4 ۶ dGo#3B ʸĴ YM3?*$o<¸Uz?Eo˾H|*C"W_mqc)acGavV#LR|+q8Ǩ nำ~ׁ?ޮ34&X:ïrV* X.J(3X4^0 _9} ՆS"Mm?4x \4mp5˖{|kQ$yBD_j8ɾtHG'اHM'8ڔOus&.ӪQ=LBt}ڏzNS?\DBRWHX\.i&dל.1H"dI e0ј`;Џޮt? qB}_ LShm P- KKiJViB-`(PlKIPH,QNT~W ()H&*4 傗9eݛa.d~\]ߔub~ޭv^_U |s7+Q| z?z_]_v1|*@_gWgZI3o^%y;}/i|ZwWW@.QHh(P2g(rE@H-9`퍧m'S;o{Y 椏ޘZs^m>56?Hm'{י`dӣq7Өd6x[VP DuT8o&-k칞p{k!XӸ bdž˺w7"7F\F==F;Hv{(,~nhBС'aBPiDHV^R. v!3$1Sް3%5R3h>LEj 4+%zɔ7w 8x6Ϧ3!:2@E.qe-3D@v3)]r$:4U@*)N٪P` m+t#(+HeX(ɑEj2mnzMu<$MLS(SsnNnѻSTK.n4*.Q NXeʲd/cuY2Aex\4sJ02DrD {^*U#2&{Q1^ԭ#;FV~NTh%dO(`nR d?RmO,gѢU6jz{h- 񚀆k{[ӍS~] ҍWj(қW=I҅jy+>ܾK<DT_wXmH_{ w`+n>ܮkK= W1$\9s09 YEWgg,& WQj0-VYKd%ph/.q!8(A{ڋ|y? _X[2N}XIzcƪ"\eF5g?ԝ\v ݐ6ϗa.فok&І8] !Aomlh[ 0_Z>}/GZD>bmDd9Y 5t3yW@ĺiZa4괶om ob)V 0Me cҴNܒCY(CfF$6= 0RW#|v`mh"'W%tF5YJl 0C-QkjR JI"d#OT<ixIwD נ7W 52*P  jQȅnAiE.w5WԆ'q05T.~?*$sQ#ORcVL_>. *F?m)muC KS35=OyT< /I#zt[Bʩ<^~@CS̪Yy )@ũANJåJ.M poi+dWL2-l6;Òd$nh4.CFa(eZ8)+`8-#K%7(ph9?gKqPTNB-1bq >$yt t2UA0>䷴XhD$cqEuY)&&AjTB"NLG˖V* %WmsJ 'u٩D%(}{F(m :, i"$>eŐQBKVB,#r4t, -)vEww2 %~}xPnZ[E"~XaRy+*J^+AsED '^[w?XI-ޝ/E=-%Y}FGDN=/T> ZAc4FǃR2 dD K6t\Z8_x S- ՂeВz;3~o3>E-)x{ =I5;pZ)$Y$+ #A!4J3FL_N㵡%kAF3\vlfqo 5#var/home/core/zuul-output/logs/kubelet.log0000644000000000000000004664662715144437567017742 0ustar rootrootFeb 15 20:16:30 crc systemd[1]: Starting Kubernetes Kubelet... Feb 15 20:16:31 crc restorecon[4580]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 15 20:16:31 crc restorecon[4580]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 15 20:16:32 crc kubenswrapper[4735]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 15 20:16:32 crc kubenswrapper[4735]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 15 20:16:32 crc kubenswrapper[4735]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 15 20:16:32 crc kubenswrapper[4735]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 15 20:16:32 crc kubenswrapper[4735]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 15 20:16:32 crc kubenswrapper[4735]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.603387 4735 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611782 4735 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611820 4735 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611831 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611840 4735 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611848 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611857 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611866 4735 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611875 4735 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611884 4735 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611892 4735 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611901 4735 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611909 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611919 4735 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611930 4735 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611940 4735 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611955 4735 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611965 4735 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.611974 4735 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612011 4735 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612023 4735 feature_gate.go:330] unrecognized feature gate: Example Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612035 4735 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612046 4735 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612056 4735 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612067 4735 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612096 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612107 4735 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612118 4735 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612128 4735 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612138 4735 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612148 4735 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612158 4735 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612168 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612178 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612188 4735 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612199 4735 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612211 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612221 4735 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612233 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612243 4735 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612253 4735 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612262 4735 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612272 4735 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612280 4735 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612288 4735 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612298 4735 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612305 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612313 4735 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612320 4735 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612328 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612336 4735 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612346 4735 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612354 4735 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612361 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612369 4735 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612377 4735 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612388 4735 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612398 4735 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612407 4735 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612414 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612422 4735 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612432 4735 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612440 4735 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612448 4735 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612455 4735 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612462 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612471 4735 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612478 4735 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612486 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612493 4735 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612500 4735 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.612508 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612751 4735 flags.go:64] FLAG: --address="0.0.0.0" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612771 4735 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612789 4735 flags.go:64] FLAG: --anonymous-auth="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612801 4735 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612813 4735 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612823 4735 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612834 4735 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612867 4735 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612877 4735 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612888 4735 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612899 4735 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612909 4735 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612919 4735 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612929 4735 flags.go:64] FLAG: --cgroup-root="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612940 4735 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.612956 4735 flags.go:64] FLAG: --client-ca-file="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613027 4735 flags.go:64] FLAG: --cloud-config="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613041 4735 flags.go:64] FLAG: --cloud-provider="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613051 4735 flags.go:64] FLAG: --cluster-dns="[]" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613065 4735 flags.go:64] FLAG: --cluster-domain="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613077 4735 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613108 4735 flags.go:64] FLAG: --config-dir="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613119 4735 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613131 4735 flags.go:64] FLAG: --container-log-max-files="5" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613147 4735 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613159 4735 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613171 4735 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613183 4735 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613193 4735 flags.go:64] FLAG: --contention-profiling="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613203 4735 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613212 4735 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613222 4735 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613231 4735 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613243 4735 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613253 4735 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613262 4735 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613272 4735 flags.go:64] FLAG: --enable-load-reader="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613281 4735 flags.go:64] FLAG: --enable-server="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613290 4735 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613307 4735 flags.go:64] FLAG: --event-burst="100" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613317 4735 flags.go:64] FLAG: --event-qps="50" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613327 4735 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613336 4735 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613346 4735 flags.go:64] FLAG: --eviction-hard="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613357 4735 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613366 4735 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613375 4735 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613385 4735 flags.go:64] FLAG: --eviction-soft="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613394 4735 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613403 4735 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613412 4735 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613421 4735 flags.go:64] FLAG: --experimental-mounter-path="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613430 4735 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613439 4735 flags.go:64] FLAG: --fail-swap-on="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613448 4735 flags.go:64] FLAG: --feature-gates="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613459 4735 flags.go:64] FLAG: --file-check-frequency="20s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613468 4735 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613478 4735 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613487 4735 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613497 4735 flags.go:64] FLAG: --healthz-port="10248" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613506 4735 flags.go:64] FLAG: --help="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613515 4735 flags.go:64] FLAG: --hostname-override="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613529 4735 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613541 4735 flags.go:64] FLAG: --http-check-frequency="20s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613552 4735 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613563 4735 flags.go:64] FLAG: --image-credential-provider-config="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613574 4735 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613586 4735 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613597 4735 flags.go:64] FLAG: --image-service-endpoint="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613608 4735 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613619 4735 flags.go:64] FLAG: --kube-api-burst="100" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613630 4735 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613642 4735 flags.go:64] FLAG: --kube-api-qps="50" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613653 4735 flags.go:64] FLAG: --kube-reserved="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613664 4735 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613675 4735 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613686 4735 flags.go:64] FLAG: --kubelet-cgroups="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613697 4735 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613708 4735 flags.go:64] FLAG: --lock-file="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613721 4735 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613735 4735 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613747 4735 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613761 4735 flags.go:64] FLAG: --log-json-split-stream="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613771 4735 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613780 4735 flags.go:64] FLAG: --log-text-split-stream="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613789 4735 flags.go:64] FLAG: --logging-format="text" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613798 4735 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613808 4735 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613817 4735 flags.go:64] FLAG: --manifest-url="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613826 4735 flags.go:64] FLAG: --manifest-url-header="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613837 4735 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613847 4735 flags.go:64] FLAG: --max-open-files="1000000" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613858 4735 flags.go:64] FLAG: --max-pods="110" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613867 4735 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613877 4735 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613886 4735 flags.go:64] FLAG: --memory-manager-policy="None" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613895 4735 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613904 4735 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613914 4735 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613924 4735 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613954 4735 flags.go:64] FLAG: --node-status-max-images="50" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613963 4735 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.613972 4735 flags.go:64] FLAG: --oom-score-adj="-999" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614019 4735 flags.go:64] FLAG: --pod-cidr="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614032 4735 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614050 4735 flags.go:64] FLAG: --pod-manifest-path="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614062 4735 flags.go:64] FLAG: --pod-max-pids="-1" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614074 4735 flags.go:64] FLAG: --pods-per-core="0" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614086 4735 flags.go:64] FLAG: --port="10250" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614098 4735 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614108 4735 flags.go:64] FLAG: --provider-id="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614116 4735 flags.go:64] FLAG: --qos-reserved="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614126 4735 flags.go:64] FLAG: --read-only-port="10255" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614136 4735 flags.go:64] FLAG: --register-node="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614146 4735 flags.go:64] FLAG: --register-schedulable="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614157 4735 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614172 4735 flags.go:64] FLAG: --registry-burst="10" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614182 4735 flags.go:64] FLAG: --registry-qps="5" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614191 4735 flags.go:64] FLAG: --reserved-cpus="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614200 4735 flags.go:64] FLAG: --reserved-memory="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614211 4735 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614220 4735 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614229 4735 flags.go:64] FLAG: --rotate-certificates="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614238 4735 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614247 4735 flags.go:64] FLAG: --runonce="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614256 4735 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614265 4735 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614275 4735 flags.go:64] FLAG: --seccomp-default="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614283 4735 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614293 4735 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614302 4735 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614311 4735 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614320 4735 flags.go:64] FLAG: --storage-driver-password="root" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614329 4735 flags.go:64] FLAG: --storage-driver-secure="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614339 4735 flags.go:64] FLAG: --storage-driver-table="stats" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614349 4735 flags.go:64] FLAG: --storage-driver-user="root" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614358 4735 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614367 4735 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614377 4735 flags.go:64] FLAG: --system-cgroups="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614386 4735 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614400 4735 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614409 4735 flags.go:64] FLAG: --tls-cert-file="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614418 4735 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614429 4735 flags.go:64] FLAG: --tls-min-version="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614439 4735 flags.go:64] FLAG: --tls-private-key-file="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614448 4735 flags.go:64] FLAG: --topology-manager-policy="none" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614458 4735 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614466 4735 flags.go:64] FLAG: --topology-manager-scope="container" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614476 4735 flags.go:64] FLAG: --v="2" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614487 4735 flags.go:64] FLAG: --version="false" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614498 4735 flags.go:64] FLAG: --vmodule="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614510 4735 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.614519 4735 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614739 4735 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614750 4735 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614759 4735 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614767 4735 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614775 4735 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614783 4735 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614792 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614800 4735 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614809 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614816 4735 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614825 4735 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614833 4735 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614840 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614848 4735 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614857 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614865 4735 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614872 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614881 4735 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614889 4735 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614896 4735 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614904 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614912 4735 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614920 4735 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614928 4735 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614936 4735 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614951 4735 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614961 4735 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.614971 4735 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615004 4735 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615013 4735 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615021 4735 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615029 4735 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615037 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615044 4735 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615053 4735 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615061 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615069 4735 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615077 4735 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615085 4735 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615093 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615102 4735 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615110 4735 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615118 4735 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615126 4735 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615135 4735 feature_gate.go:330] unrecognized feature gate: Example Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615142 4735 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615150 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615161 4735 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615170 4735 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615178 4735 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615187 4735 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615195 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615203 4735 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615211 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615233 4735 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615243 4735 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615259 4735 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615268 4735 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615277 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615286 4735 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615294 4735 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615302 4735 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615310 4735 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615318 4735 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615326 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615334 4735 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615344 4735 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615353 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615362 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615372 4735 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.615383 4735 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.615407 4735 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.627867 4735 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.627928 4735 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628136 4735 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628161 4735 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628172 4735 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628182 4735 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628191 4735 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628200 4735 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628209 4735 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628217 4735 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628225 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628233 4735 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628240 4735 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628248 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628256 4735 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628263 4735 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628271 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628279 4735 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628286 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628295 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628306 4735 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628319 4735 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628329 4735 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628339 4735 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628347 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628356 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628365 4735 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628373 4735 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628381 4735 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628389 4735 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628397 4735 feature_gate.go:330] unrecognized feature gate: Example Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628406 4735 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628414 4735 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628423 4735 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628432 4735 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628441 4735 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628450 4735 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628458 4735 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628466 4735 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628474 4735 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628485 4735 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628494 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628501 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628509 4735 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628516 4735 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628524 4735 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628533 4735 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628541 4735 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628549 4735 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628556 4735 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628564 4735 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628572 4735 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628579 4735 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628587 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628594 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628602 4735 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628610 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628617 4735 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628625 4735 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628633 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628640 4735 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628651 4735 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628660 4735 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628669 4735 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628678 4735 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628687 4735 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628695 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628703 4735 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628711 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628719 4735 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628726 4735 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628734 4735 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.628744 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.628757 4735 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629041 4735 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629056 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629067 4735 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629075 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629082 4735 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629090 4735 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629098 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629106 4735 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629117 4735 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629126 4735 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629133 4735 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629141 4735 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629149 4735 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629157 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629164 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629172 4735 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629181 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629189 4735 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629197 4735 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629204 4735 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629212 4735 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629219 4735 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629257 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629267 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629275 4735 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629283 4735 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629292 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629300 4735 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629308 4735 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629317 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629325 4735 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629333 4735 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629341 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629349 4735 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629360 4735 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629368 4735 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629376 4735 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629385 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629393 4735 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629401 4735 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629410 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629418 4735 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629426 4735 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629434 4735 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629442 4735 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629451 4735 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629459 4735 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629470 4735 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629483 4735 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629494 4735 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629503 4735 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629511 4735 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629520 4735 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629530 4735 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629537 4735 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629548 4735 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629559 4735 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629569 4735 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629579 4735 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629587 4735 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629595 4735 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629604 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629612 4735 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629621 4735 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629629 4735 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629637 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629645 4735 feature_gate.go:330] unrecognized feature gate: Example Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629653 4735 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629661 4735 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629670 4735 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.629679 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.629692 4735 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.630013 4735 server.go:940] "Client rotation is on, will bootstrap in background" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.637456 4735 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.637615 4735 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.639429 4735 server.go:997] "Starting client certificate rotation" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.639478 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.639729 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-09 06:29:37.307109958 +0000 UTC Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.639917 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.669430 4735 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.672287 4735 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.673481 4735 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.687404 4735 log.go:25] "Validated CRI v1 runtime API" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.731096 4735 log.go:25] "Validated CRI v1 image API" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.733360 4735 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.739648 4735 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-15-20-08-37-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.739734 4735 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.763724 4735 manager.go:217] Machine: {Timestamp:2026-02-15 20:16:32.759108924 +0000 UTC m=+0.625124627 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199472640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:da2405e5-a6ff-4f93-8eee-d5a8c6845038 BootID:b29648d8-57a5-4453-a2ae-0d1a863f2ae4 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076107 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599734272 Type:vfs Inodes:3076107 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039894528 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:aa:a7:4e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:aa:a7:4e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:31:b5:0f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:83:d3:4d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ec:88:ed Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e9:f6:4a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:96:5b:3b:05:a3:be Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:4e:e4:f2:32:04:d2 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199472640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.764124 4735 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.764348 4735 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.766908 4735 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.767247 4735 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.767308 4735 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.767624 4735 topology_manager.go:138] "Creating topology manager with none policy" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.767643 4735 container_manager_linux.go:303] "Creating device plugin manager" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.768530 4735 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.768605 4735 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.768845 4735 state_mem.go:36] "Initialized new in-memory state store" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.769029 4735 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.773365 4735 kubelet.go:418] "Attempting to sync node with API server" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.773405 4735 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.773481 4735 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.773506 4735 kubelet.go:324] "Adding apiserver pod source" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.773527 4735 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.779616 4735 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.781644 4735 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.781565 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.781751 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.781583 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.781813 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.784260 4735 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.785867 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.785925 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.785954 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.786007 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.786034 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.786080 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.786095 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.786119 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.786135 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.786150 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.786185 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.786200 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.790184 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.790966 4735 server.go:1280] "Started kubelet" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.796414 4735 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.796506 4735 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 15 20:16:32 crc systemd[1]: Started Kubernetes Kubelet. Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.797732 4735 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.797470 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.802031 4735 server.go:460] "Adding debug handlers to kubelet server" Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.800592 4735 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.82:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189484d26e3cb8de default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-15 20:16:32.79091939 +0000 UTC m=+0.656935053,LastTimestamp:2026-02-15 20:16:32.79091939 +0000 UTC m=+0.656935053,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.806764 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.808144 4735 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.808160 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 23:25:08.239994475 +0000 UTC Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.809224 4735 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.809411 4735 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.809693 4735 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.809423 4735 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.810182 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.810287 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817172 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817230 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817243 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817256 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817270 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817284 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817297 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817310 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817325 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817339 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817351 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817364 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817389 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817403 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817417 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817430 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817445 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817464 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817492 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817505 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817518 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817531 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817544 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817558 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817576 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817604 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817623 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817639 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817654 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817666 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817679 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817693 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.817667 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="200ms" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817706 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817752 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817787 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817810 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817833 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817853 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817872 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817892 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817912 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.817935 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.818605 4735 factory.go:55] Registering systemd factory Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.818639 4735 factory.go:221] Registration of the systemd container factory successfully Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821366 4735 factory.go:153] Registering CRI-O factory Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821401 4735 factory.go:221] Registration of the crio container factory successfully Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821486 4735 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821508 4735 factory.go:103] Registering Raw factory Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821526 4735 manager.go:1196] Started watching for new ooms in manager Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821568 4735 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821617 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821650 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821679 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821704 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821723 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821743 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821764 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821785 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821806 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821827 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821858 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821879 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821906 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821931 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.821959 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822010 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822034 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822055 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822077 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822098 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822121 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822143 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822164 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822186 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822206 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822225 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822248 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822269 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822289 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822310 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822331 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822354 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822377 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822400 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822422 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822443 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822463 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822484 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822509 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822540 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822562 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822587 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822607 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822647 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822669 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822690 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822710 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822729 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822750 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822770 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822790 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822811 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822832 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822853 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822875 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822896 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822917 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822940 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822968 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823012 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823072 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823094 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823124 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823148 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823172 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823194 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823218 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823242 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823264 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823287 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823309 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823330 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823353 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823373 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823395 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823413 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823432 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823454 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823474 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823495 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823517 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823536 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823557 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823578 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823600 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823621 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823642 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823663 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823683 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823706 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823726 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823748 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823769 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823794 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823814 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823841 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823937 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.823960 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824053 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824078 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824100 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824120 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824139 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824160 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824179 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824199 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824221 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824242 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824265 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824286 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824308 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824329 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824351 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824372 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824391 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824410 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824430 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824449 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824468 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824488 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824507 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824537 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824557 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824578 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824598 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824619 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824639 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824660 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824683 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824702 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824723 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824744 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824765 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824784 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824805 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824824 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824863 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824883 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824903 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.824923 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825094 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825117 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825141 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825160 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825184 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825203 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825225 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825245 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825265 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825286 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825306 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825327 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825346 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825370 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825390 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825410 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825429 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825449 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825474 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825492 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825514 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825535 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825555 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825577 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825597 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825616 4735 reconstruct.go:97] "Volume reconstruction finished" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.825629 4735 reconciler.go:26] "Reconciler: start to sync state" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.822105 4735 manager.go:319] Starting recovery of all containers Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.866135 4735 manager.go:324] Recovery completed Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.881468 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.882083 4735 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.883918 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.883967 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.884006 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.884724 4735 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.884744 4735 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.884781 4735 state_mem.go:36] "Initialized new in-memory state store" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.885064 4735 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.885472 4735 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.885513 4735 kubelet.go:2335] "Starting kubelet main sync loop" Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.885616 4735 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 15 20:16:32 crc kubenswrapper[4735]: W0215 20:16:32.886108 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.886179 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.909961 4735 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.912928 4735 policy_none.go:49] "None policy: Start" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.913905 4735 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.913933 4735 state_mem.go:35] "Initializing new in-memory state store" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.971867 4735 manager.go:334] "Starting Device Plugin manager" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.973090 4735 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.973127 4735 server.go:79] "Starting device plugin registration server" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.973802 4735 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.973833 4735 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.974185 4735 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.974466 4735 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.974496 4735 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 15 20:16:32 crc kubenswrapper[4735]: E0215 20:16:32.985154 4735 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.986530 4735 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.986666 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.988120 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.988174 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.988198 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.988488 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.989556 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.989638 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.989920 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.989991 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.990009 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.990201 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.990348 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.990398 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.991093 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.991121 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.991136 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.991154 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.991200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.991226 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.991487 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.991656 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.991730 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.992925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.992925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993037 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993119 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993079 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993171 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993185 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993245 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993491 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993657 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.993724 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.994817 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.994879 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.994917 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.994989 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.995020 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.995037 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.995353 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.995437 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.996678 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.996746 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:32 crc kubenswrapper[4735]: I0215 20:16:32.996771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:33 crc kubenswrapper[4735]: E0215 20:16:33.018602 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="400ms" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029034 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029166 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029215 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029252 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029289 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029487 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029563 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029609 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029641 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029673 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029715 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029778 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.029903 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.030016 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.030071 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.077918 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.079355 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.079411 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.079429 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.079466 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 15 20:16:33 crc kubenswrapper[4735]: E0215 20:16:33.080126 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.82:6443: connect: connection refused" node="crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.131650 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.131714 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.131753 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.131783 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.131815 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.131846 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.131877 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.131906 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.131936 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132009 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132045 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132077 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132108 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132136 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132164 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132633 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132677 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132707 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132753 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132797 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132842 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132735 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132845 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132846 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132867 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132873 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132901 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132902 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132908 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.132936 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.281163 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.282857 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.282920 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.282940 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.283023 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 15 20:16:33 crc kubenswrapper[4735]: E0215 20:16:33.283745 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.82:6443: connect: connection refused" node="crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.340502 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.364095 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.372021 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.397109 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.402619 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:33 crc kubenswrapper[4735]: W0215 20:16:33.407491 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-3953f07f5848441184f3aa0db4dd22054bcb10bb902c09812b27f5a55859c0a2 WatchSource:0}: Error finding container 3953f07f5848441184f3aa0db4dd22054bcb10bb902c09812b27f5a55859c0a2: Status 404 returned error can't find the container with id 3953f07f5848441184f3aa0db4dd22054bcb10bb902c09812b27f5a55859c0a2 Feb 15 20:16:33 crc kubenswrapper[4735]: W0215 20:16:33.409250 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-470406d28fa1a3a53d41d9ea40fd8c1d16f3f28bd16f7de7add53fae6dfa1285 WatchSource:0}: Error finding container 470406d28fa1a3a53d41d9ea40fd8c1d16f3f28bd16f7de7add53fae6dfa1285: Status 404 returned error can't find the container with id 470406d28fa1a3a53d41d9ea40fd8c1d16f3f28bd16f7de7add53fae6dfa1285 Feb 15 20:16:33 crc kubenswrapper[4735]: E0215 20:16:33.421629 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="800ms" Feb 15 20:16:33 crc kubenswrapper[4735]: W0215 20:16:33.431786 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-ea86dc4400324d460d3f499d33511a6442addc5379dc46595ad9019cda80c1d7 WatchSource:0}: Error finding container ea86dc4400324d460d3f499d33511a6442addc5379dc46595ad9019cda80c1d7: Status 404 returned error can't find the container with id ea86dc4400324d460d3f499d33511a6442addc5379dc46595ad9019cda80c1d7 Feb 15 20:16:33 crc kubenswrapper[4735]: W0215 20:16:33.602646 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:33 crc kubenswrapper[4735]: E0215 20:16:33.602762 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.684080 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.685644 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.685715 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.685737 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.685778 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 15 20:16:33 crc kubenswrapper[4735]: E0215 20:16:33.686372 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.82:6443: connect: connection refused" node="crc" Feb 15 20:16:33 crc kubenswrapper[4735]: W0215 20:16:33.725290 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:33 crc kubenswrapper[4735]: E0215 20:16:33.725377 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:33 crc kubenswrapper[4735]: W0215 20:16:33.773202 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:33 crc kubenswrapper[4735]: E0215 20:16:33.773324 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.802441 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.808522 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 02:18:27.595063635 +0000 UTC Feb 15 20:16:33 crc kubenswrapper[4735]: E0215 20:16:33.885256 4735 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.82:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189484d26e3cb8de default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-15 20:16:32.79091939 +0000 UTC m=+0.656935053,LastTimestamp:2026-02-15 20:16:32.79091939 +0000 UTC m=+0.656935053,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.892648 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"470406d28fa1a3a53d41d9ea40fd8c1d16f3f28bd16f7de7add53fae6dfa1285"} Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.894856 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3953f07f5848441184f3aa0db4dd22054bcb10bb902c09812b27f5a55859c0a2"} Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.897062 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ea86dc4400324d460d3f499d33511a6442addc5379dc46595ad9019cda80c1d7"} Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.898495 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"88314aca070e0c4e94fb136774950c561aecc08185b53e472ae506626dc6bf86"} Feb 15 20:16:33 crc kubenswrapper[4735]: I0215 20:16:33.900514 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6f9600876648c07b36b11a4f5d954f22fd9f0c9548105a12ce5d143591e768c6"} Feb 15 20:16:34 crc kubenswrapper[4735]: W0215 20:16:34.063371 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:34 crc kubenswrapper[4735]: E0215 20:16:34.064038 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:34 crc kubenswrapper[4735]: E0215 20:16:34.223394 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="1.6s" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.487468 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.492672 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.492751 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.492772 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.492817 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 15 20:16:34 crc kubenswrapper[4735]: E0215 20:16:34.493538 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.82:6443: connect: connection refused" node="crc" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.802802 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.809251 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 10:20:06.907789571 +0000 UTC Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.837607 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 15 20:16:34 crc kubenswrapper[4735]: E0215 20:16:34.839155 4735 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.908527 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33"} Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.908614 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f"} Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.908629 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67"} Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.908643 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063"} Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.908771 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.909972 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.910007 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.910018 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.912860 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934" exitCode=0 Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.912939 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934"} Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.913073 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.914351 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.914388 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.914402 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.916396 4735 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="97918334d9d8c784029961323c130538e12945a02fe644a5a44d206ea6120bf1" exitCode=0 Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.916484 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"97918334d9d8c784029961323c130538e12945a02fe644a5a44d206ea6120bf1"} Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.916579 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.917400 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.918289 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.918335 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.918350 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.918866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.918916 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.918937 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.921675 4735 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8f46579da97407cf36a3e436f4f988a11e5cd953b8202e495eb6c34b55fa24dc" exitCode=0 Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.921863 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.921862 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8f46579da97407cf36a3e436f4f988a11e5cd953b8202e495eb6c34b55fa24dc"} Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.925324 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.925367 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.925380 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.931773 4735 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433" exitCode=0 Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.931833 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433"} Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.931891 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.935446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.935501 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:34 crc kubenswrapper[4735]: I0215 20:16:34.935514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:35 crc kubenswrapper[4735]: W0215 20:16:35.793314 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:35 crc kubenswrapper[4735]: E0215 20:16:35.793443 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.82:6443: connect: connection refused" logger="UnhandledError" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.802372 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.82:6443: connect: connection refused Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.809503 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 07:57:40.183115624 +0000 UTC Feb 15 20:16:35 crc kubenswrapper[4735]: E0215 20:16:35.824235 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="3.2s" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.937690 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de"} Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.937742 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3"} Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.937758 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392"} Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.940823 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ef07321fa57ad8cbabfff7062bc63516c964e96f75fbd6409dc53c63a476a228"} Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.940896 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.942463 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.942493 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.942502 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.944494 4735 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2cee99e1ce616804edc68f527894ce6df24b999bf9986cb047b519ba2134d731" exitCode=0 Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.944594 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2cee99e1ce616804edc68f527894ce6df24b999bf9986cb047b519ba2134d731"} Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.944620 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.945743 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.945783 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.945795 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.950458 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.950556 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"740e7c6d10a65b9d2bdc9cba9d431aabb9c20e9bcce1fb2f40366a8a53a4f931"} Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.950591 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.950595 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"aa568691661d11a2d6c03d89986f384f64101f0a2047d4865cc4a4b04fc2483c"} Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.951053 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4fc54fa7126011e839c783514b4d553d25a86a4175c053e84301561b281d19fd"} Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.952300 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.952343 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.952360 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.952512 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.952533 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:35 crc kubenswrapper[4735]: I0215 20:16:35.952543 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.094182 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.095569 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.095622 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.095641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.095690 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 15 20:16:36 crc kubenswrapper[4735]: E0215 20:16:36.096287 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.82:6443: connect: connection refused" node="crc" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.810245 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 22:34:20.784823324 +0000 UTC Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.955558 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad"} Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.955609 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b"} Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.955753 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.957692 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.957738 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.957756 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.959929 4735 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ca03f802ccc6cde403be4b0330beec114f4b3226ac032a574c7ee4a8dcda117e" exitCode=0 Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.960042 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.960079 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.960595 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ca03f802ccc6cde403be4b0330beec114f4b3226ac032a574c7ee4a8dcda117e"} Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.960650 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.960755 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.961203 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.961227 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.961241 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.961720 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.961753 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.961765 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.961839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.961860 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:36 crc kubenswrapper[4735]: I0215 20:16:36.961870 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.313862 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.810383 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 00:25:48.9431456 +0000 UTC Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.966585 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.967105 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"04d68798fbaab0dc2a331e7fafd04278941907ec0e60f502caf28bd71288d485"} Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.967157 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"787741fc3d1a43701dd8b94569a769a8b7d9d5617e7c63d6e2ca978348af12d5"} Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.967178 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7c08e06c9a8cbdf6dceec31a1a9a0da6be4b2b9fa0e95f2e08b7592afff4a01b"} Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.967218 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.967268 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.967708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.967777 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.967809 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.968000 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.968042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:37 crc kubenswrapper[4735]: I0215 20:16:37.968053 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:38 crc kubenswrapper[4735]: I0215 20:16:38.811124 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 17:11:18.797182902 +0000 UTC Feb 15 20:16:38 crc kubenswrapper[4735]: I0215 20:16:38.975870 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bd2900efa729767ae2f2c2919f83b2bcdd560bf7ae52f5bd08156f31afe890dc"} Feb 15 20:16:38 crc kubenswrapper[4735]: I0215 20:16:38.975984 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"82d77383029a74a6b51bf089b59c85ef643a2aa11335e89c491ed00c05fb7a7d"} Feb 15 20:16:38 crc kubenswrapper[4735]: I0215 20:16:38.976162 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:38 crc kubenswrapper[4735]: I0215 20:16:38.977491 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:38 crc kubenswrapper[4735]: I0215 20:16:38.977536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:38 crc kubenswrapper[4735]: I0215 20:16:38.977549 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.042067 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.297358 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.298932 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.299018 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.299035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.299069 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.811618 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 07:47:57.218164742 +0000 UTC Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.977901 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.978941 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.978993 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:39 crc kubenswrapper[4735]: I0215 20:16:39.979004 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:40 crc kubenswrapper[4735]: I0215 20:16:40.707592 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 15 20:16:40 crc kubenswrapper[4735]: I0215 20:16:40.812789 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 19:18:52.442237297 +0000 UTC Feb 15 20:16:40 crc kubenswrapper[4735]: I0215 20:16:40.981117 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:40 crc kubenswrapper[4735]: I0215 20:16:40.983510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:40 crc kubenswrapper[4735]: I0215 20:16:40.983566 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:40 crc kubenswrapper[4735]: I0215 20:16:40.983584 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.032203 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.032458 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.032532 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.034514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.034564 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.034581 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.155716 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.156423 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.158140 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.158196 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.158214 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.813543 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 13:22:37.007750692 +0000 UTC Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.935988 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.984759 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.985553 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.987413 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.987474 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:41 crc kubenswrapper[4735]: I0215 20:16:41.987497 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.462166 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.815005 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 12:08:41.894515435 +0000 UTC Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.854632 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.855203 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.856666 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.856751 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.856769 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.862269 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:42 crc kubenswrapper[4735]: E0215 20:16:42.985347 4735 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.987098 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.987139 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.987240 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.992228 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.992466 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.992303 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.992656 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.992677 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:42 crc kubenswrapper[4735]: I0215 20:16:42.992618 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.816459 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 07:09:45.893904303 +0000 UTC Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.970597 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.970914 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.972571 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.972634 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.972652 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.989472 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.990644 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.990812 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:43 crc kubenswrapper[4735]: I0215 20:16:43.990981 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:44 crc kubenswrapper[4735]: I0215 20:16:44.047561 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:44 crc kubenswrapper[4735]: I0215 20:16:44.354887 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:44 crc kubenswrapper[4735]: I0215 20:16:44.817160 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 21:58:48.645460651 +0000 UTC Feb 15 20:16:44 crc kubenswrapper[4735]: I0215 20:16:44.992281 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:44 crc kubenswrapper[4735]: I0215 20:16:44.993763 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:44 crc kubenswrapper[4735]: I0215 20:16:44.993834 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:44 crc kubenswrapper[4735]: I0215 20:16:44.993858 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:45 crc kubenswrapper[4735]: I0215 20:16:45.817630 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 20:04:00.211523232 +0000 UTC Feb 15 20:16:45 crc kubenswrapper[4735]: I0215 20:16:45.995744 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:45 crc kubenswrapper[4735]: I0215 20:16:45.997401 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:45 crc kubenswrapper[4735]: I0215 20:16:45.997464 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:45 crc kubenswrapper[4735]: I0215 20:16:45.997482 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:46 crc kubenswrapper[4735]: W0215 20:16:46.630614 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 15 20:16:46 crc kubenswrapper[4735]: I0215 20:16:46.630742 4735 trace.go:236] Trace[293620313]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Feb-2026 20:16:36.628) (total time: 10001ms): Feb 15 20:16:46 crc kubenswrapper[4735]: Trace[293620313]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (20:16:46.630) Feb 15 20:16:46 crc kubenswrapper[4735]: Trace[293620313]: [10.001903846s] [10.001903846s] END Feb 15 20:16:46 crc kubenswrapper[4735]: E0215 20:16:46.630776 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 15 20:16:46 crc kubenswrapper[4735]: I0215 20:16:46.803159 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 15 20:16:46 crc kubenswrapper[4735]: W0215 20:16:46.813040 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 15 20:16:46 crc kubenswrapper[4735]: I0215 20:16:46.813209 4735 trace.go:236] Trace[1064383817]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Feb-2026 20:16:36.811) (total time: 10001ms): Feb 15 20:16:46 crc kubenswrapper[4735]: Trace[1064383817]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (20:16:46.813) Feb 15 20:16:46 crc kubenswrapper[4735]: Trace[1064383817]: [10.001834503s] [10.001834503s] END Feb 15 20:16:46 crc kubenswrapper[4735]: E0215 20:16:46.813259 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 15 20:16:46 crc kubenswrapper[4735]: I0215 20:16:46.818805 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 22:25:44.372225856 +0000 UTC Feb 15 20:16:46 crc kubenswrapper[4735]: I0215 20:16:46.999656 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.001511 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b" exitCode=255 Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.001557 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b"} Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.001692 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.002446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.002474 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.002482 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.002898 4735 scope.go:117] "RemoveContainer" containerID="45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b" Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.047389 4735 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.047478 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 15 20:16:47 crc kubenswrapper[4735]: W0215 20:16:47.243541 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.243680 4735 trace.go:236] Trace[1200469977]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Feb-2026 20:16:37.242) (total time: 10000ms): Feb 15 20:16:47 crc kubenswrapper[4735]: Trace[1200469977]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (20:16:47.243) Feb 15 20:16:47 crc kubenswrapper[4735]: Trace[1200469977]: [10.000953367s] [10.000953367s] END Feb 15 20:16:47 crc kubenswrapper[4735]: E0215 20:16:47.243727 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.819465 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 01:38:10.573785926 +0000 UTC Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.836921 4735 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.837000 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.844122 4735 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 15 20:16:47 crc kubenswrapper[4735]: I0215 20:16:47.844163 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 15 20:16:48 crc kubenswrapper[4735]: I0215 20:16:48.005989 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 15 20:16:48 crc kubenswrapper[4735]: I0215 20:16:48.008289 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e"} Feb 15 20:16:48 crc kubenswrapper[4735]: I0215 20:16:48.008468 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:48 crc kubenswrapper[4735]: I0215 20:16:48.009394 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:48 crc kubenswrapper[4735]: I0215 20:16:48.009424 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:48 crc kubenswrapper[4735]: I0215 20:16:48.009435 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:48 crc kubenswrapper[4735]: I0215 20:16:48.820596 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 20:43:24.374805908 +0000 UTC Feb 15 20:16:49 crc kubenswrapper[4735]: I0215 20:16:49.821558 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 13:35:51.938017688 +0000 UTC Feb 15 20:16:50 crc kubenswrapper[4735]: I0215 20:16:50.131254 4735 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 15 20:16:50 crc kubenswrapper[4735]: I0215 20:16:50.756360 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 15 20:16:50 crc kubenswrapper[4735]: I0215 20:16:50.756625 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:50 crc kubenswrapper[4735]: I0215 20:16:50.758544 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:50 crc kubenswrapper[4735]: I0215 20:16:50.758612 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:50 crc kubenswrapper[4735]: I0215 20:16:50.758632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:50 crc kubenswrapper[4735]: I0215 20:16:50.771764 4735 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 15 20:16:50 crc kubenswrapper[4735]: I0215 20:16:50.777737 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 15 20:16:50 crc kubenswrapper[4735]: I0215 20:16:50.821693 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 10:36:50.439774911 +0000 UTC Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.016419 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.017850 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.017923 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.017942 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.629144 4735 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.782930 4735 apiserver.go:52] "Watching apiserver" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.790076 4735 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.790423 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.791140 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.791165 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.791379 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.791480 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.791629 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 15 20:16:51 crc kubenswrapper[4735]: E0215 20:16:51.791634 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:16:51 crc kubenswrapper[4735]: E0215 20:16:51.791851 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.792554 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:51 crc kubenswrapper[4735]: E0215 20:16:51.792746 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.796152 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.796649 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.796844 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.801057 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.803100 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.803166 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.803387 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.803632 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.803636 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.811261 4735 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.822296 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 20:51:46.30997025 +0000 UTC Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.848245 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.868860 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.886083 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.906023 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.922030 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.934233 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.934779 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.939521 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.942790 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.959666 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.959881 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.976355 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:51 crc kubenswrapper[4735]: I0215 20:16:51.987630 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.000201 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.011869 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.024886 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.040940 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.056118 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.813035 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.815899 4735 trace.go:236] Trace[1688668682]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Feb-2026 20:16:39.276) (total time: 13539ms): Feb 15 20:16:52 crc kubenswrapper[4735]: Trace[1688668682]: ---"Objects listed" error: 13539ms (20:16:52.815) Feb 15 20:16:52 crc kubenswrapper[4735]: Trace[1688668682]: [13.539796557s] [13.539796557s] END Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.815992 4735 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.818797 4735 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.819867 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.822730 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 07:28:32.514810297 +0000 UTC Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.851234 4735 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.900740 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.919888 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.920536 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.920777 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.920991 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.921187 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.921367 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.921567 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.921742 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.921911 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.922049 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.922229 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.922358 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.922548 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.922732 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.922910 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.923142 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.923323 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.923505 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.923720 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.923923 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.924193 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.924389 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.924561 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.924731 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925349 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925448 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925498 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925542 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925583 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925618 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925664 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925708 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925750 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925792 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925845 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925896 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925933 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925997 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.926043 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.926079 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.926113 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.926151 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.926187 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.926227 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.926263 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.922600 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.923098 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.923302 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.923495 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.923681 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.924370 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.924594 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925516 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925700 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925422 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.925936 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.926219 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.926437 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.927008 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.927191 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.927528 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.928087 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.928281 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.928462 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.928675 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.928858 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929581 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929661 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929707 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929755 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929800 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929840 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929887 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929931 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930004 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930047 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930092 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930151 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930193 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930236 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930282 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930320 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930369 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930414 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930457 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930492 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930537 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930580 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930616 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930658 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930704 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930900 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930981 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931027 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931065 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931107 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931151 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931187 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929118 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929632 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.929843 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930239 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.930830 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931501 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931787 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.933338 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931932 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931927 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.932476 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.932747 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.932968 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.932632 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.935164 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.935312 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.936066 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.936278 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:16:53.43625283 +0000 UTC m=+21.302268453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940077 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940102 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940151 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.936398 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940246 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.936755 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.937041 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.937263 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.937965 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.937973 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940373 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.937667 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.938803 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.939251 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.939570 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.938580 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.939794 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940716 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940748 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940853 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940908 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.940932 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.941334 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.941339 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.941631 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.941885 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.942494 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.942611 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.942761 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943123 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.931227 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943499 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943544 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943641 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943676 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943706 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943704 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943738 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943766 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943796 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943823 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943853 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943885 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943912 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943937 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.943986 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944013 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944037 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944063 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944066 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944088 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944177 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944219 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944245 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944315 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944323 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944348 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944377 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944405 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944433 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944463 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944488 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944531 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944536 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944600 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944633 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944660 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944684 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944710 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944737 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944761 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.944993 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945023 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945049 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945076 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945103 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945129 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945155 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945235 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945316 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945348 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945370 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945396 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945457 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945485 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945506 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945531 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945555 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945577 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945605 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945630 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945688 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945721 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945798 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945823 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945853 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945880 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945905 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.946332 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.946815 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.945926 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.947023 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.947131 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.947154 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.947219 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.947266 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.947285 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.948331 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.948474 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.948542 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.948591 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.948633 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.948679 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.948718 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.947722 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.949121 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.949265 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.949280 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.949294 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.949324 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.949615 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.949724 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.949899 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950039 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950080 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950103 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950123 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950144 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950165 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950186 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950200 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950205 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950272 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950305 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950335 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950361 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950519 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950503 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950552 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950569 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.950914 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.951903 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.952668 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.953001 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.953163 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.953968 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954224 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954414 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954604 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954659 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954723 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954798 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954833 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954873 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954883 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954908 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.954995 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.955020 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.955056 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.955080 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.955230 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.955527 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.955748 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.955724 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.955778 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.956110 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.956752 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.957370 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.957390 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.952005 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.958465 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.958739 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.958785 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.958836 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.958902 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.957296 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959056 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.958880 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959468 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959526 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959168 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959231 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959220 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959334 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.958490 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959718 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959755 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959785 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959798 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.959816 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960032 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960091 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960132 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960176 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960215 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960258 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960327 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960415 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960670 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960790 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960826 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.960209 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.961096 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.961106 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.961130 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.961304 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.961598 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.961745 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.961761 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.961795 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.962174 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.962168 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.962510 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.962622 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963161 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963205 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963410 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963560 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963657 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963756 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963994 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964243 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964305 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964349 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964421 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964464 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964506 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964551 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964595 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964639 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964750 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964805 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964884 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964926 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.967778 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969585 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969630 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969746 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969758 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969769 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969778 4735 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969790 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969801 4735 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969811 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969820 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969830 4735 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969840 4735 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969849 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969859 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969869 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969879 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969888 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969898 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969908 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.971018 4735 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.972172 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.972726 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.972751 4735 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.972909 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.972933 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973052 4735 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973074 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973092 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973185 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973204 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973224 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973361 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973384 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973402 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973449 4735 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973467 4735 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973487 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973531 4735 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973551 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973566 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973582 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973628 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973648 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973666 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973709 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973726 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973743 4735 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973759 4735 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973796 4735 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973815 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973832 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973874 4735 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973896 4735 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973912 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.973928 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974026 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974041 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974058 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974139 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974163 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974182 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974232 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974251 4735 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974263 4735 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974275 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974309 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974321 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974332 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974344 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974356 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974387 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974401 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974414 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974426 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974440 4735 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974475 4735 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974492 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974508 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974524 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974568 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974584 4735 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974599 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974616 4735 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974670 4735 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974683 4735 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974695 4735 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974706 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974717 4735 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974749 4735 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974764 4735 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974785 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974799 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974831 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974843 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974854 4735 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974868 4735 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974880 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974910 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974921 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974933 4735 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974968 4735 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974980 4735 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.974993 4735 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975006 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975018 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975050 4735 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975062 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975073 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975085 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975096 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975124 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975138 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975150 4735 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975221 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975235 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975252 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975264 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975305 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975328 4735 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975341 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975354 4735 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975386 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975398 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975410 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975421 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.975434 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976061 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976418 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976457 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976578 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976592 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976607 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976619 4735 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976655 4735 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976667 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976678 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976691 4735 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976703 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976736 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976752 4735 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976770 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976787 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976840 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976857 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976872 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.976887 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963514 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963529 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.963777 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.964781 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.965789 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.966193 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.966661 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.978478 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.967069 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.968127 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.968233 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969149 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969251 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969849 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.970456 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.970488 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.970517 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.970840 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.971265 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.971326 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.978656 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:53.478630223 +0000 UTC m=+21.344645866 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.978763 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:53.478753847 +0000 UTC m=+21.344769490 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.969133 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.990585 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.995154 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.995183 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.995197 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:52 crc kubenswrapper[4735]: E0215 20:16:52.995265 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:53.49524436 +0000 UTC m=+21.361259983 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.999423 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.999688 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:52 crc kubenswrapper[4735]: I0215 20:16:52.999920 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.000078 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.000242 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.000549 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.000838 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.002622 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.004245 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.004319 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.004419 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.004983 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.005063 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.007188 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.008281 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.014087 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.020081 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.020193 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.020512 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.027748 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.028018 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.054629 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.054664 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.054676 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.054734 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:53.554711082 +0000 UTC m=+21.420726705 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.055133 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.055166 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.055449 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.056241 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.056498 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.057549 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.064460 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077523 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077569 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077639 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077652 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077661 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077671 4735 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077683 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077694 4735 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077704 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077713 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077722 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077732 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077740 4735 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077748 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077756 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077764 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077772 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077781 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077789 4735 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077798 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077807 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077817 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077825 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077835 4735 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077845 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077855 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077865 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077875 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077884 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077893 4735 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077902 4735 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077910 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077919 4735 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077928 4735 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077938 4735 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077960 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077969 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077978 4735 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077987 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.077997 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.078007 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.078015 4735 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.078055 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.078184 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.081357 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.081370 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.081660 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.081937 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.082058 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.082432 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.082590 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.083290 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.083301 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.083605 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.083597 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.090169 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.093183 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.104178 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.108812 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178420 4735 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178449 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178458 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178468 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178479 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178487 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178497 4735 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178506 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178514 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178528 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178536 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.178544 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.317293 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.327370 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 15 20:16:53 crc kubenswrapper[4735]: W0215 20:16:53.327971 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-11114c3da136edb614306b8dbe705f8dacc2207daf6c1e4bf50fa7598d84ed19 WatchSource:0}: Error finding container 11114c3da136edb614306b8dbe705f8dacc2207daf6c1e4bf50fa7598d84ed19: Status 404 returned error can't find the container with id 11114c3da136edb614306b8dbe705f8dacc2207daf6c1e4bf50fa7598d84ed19 Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.332914 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 15 20:16:53 crc kubenswrapper[4735]: W0215 20:16:53.346400 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-ebf90d3b40d1da624c0e16611b2310d2929d586bba6b98813cf1904ac69dfdbb WatchSource:0}: Error finding container ebf90d3b40d1da624c0e16611b2310d2929d586bba6b98813cf1904ac69dfdbb: Status 404 returned error can't find the container with id ebf90d3b40d1da624c0e16611b2310d2929d586bba6b98813cf1904ac69dfdbb Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.481252 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.481324 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.481373 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.481394 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:16:54.481373329 +0000 UTC m=+22.347388942 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.481452 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.481477 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.481507 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:54.481495532 +0000 UTC m=+22.347511155 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.481519 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:54.481513483 +0000 UTC m=+22.347529106 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.582373 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.582412 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.582537 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.582551 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.582562 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.582602 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:54.582589088 +0000 UTC m=+22.448604701 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.582899 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.583135 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.583142 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.583166 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:54.583159145 +0000 UTC m=+22.449174768 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.823616 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 03:18:57.406224565 +0000 UTC Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.886448 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.886626 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.887060 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.887123 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:16:53 crc kubenswrapper[4735]: I0215 20:16:53.887170 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:16:53 crc kubenswrapper[4735]: E0215 20:16:53.887213 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.040379 4735 csr.go:261] certificate signing request csr-rrj4v is approved, waiting to be issued Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.051523 4735 csr.go:257] certificate signing request csr-rrj4v is issued Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.058976 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654"} Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.059027 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a"} Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.059042 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ebf90d3b40d1da624c0e16611b2310d2929d586bba6b98813cf1904ac69dfdbb"} Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.060754 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"11114c3da136edb614306b8dbe705f8dacc2207daf6c1e4bf50fa7598d84ed19"} Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.062211 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397"} Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.062262 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3fa3f96d5e2a61a521305049c9d43bde242aa2deafdf6d67764a85be351ab048"} Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.064076 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.074162 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.076237 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.092600 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.095819 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.124112 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.167091 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.198470 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.211350 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.234624 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.274065 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.297796 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.319404 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.344674 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.362688 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.378793 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.391910 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.412212 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.433810 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-sdcbj"] Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.434148 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-kjps6"] Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.434334 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-sdcbj" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.434418 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.436932 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.436957 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.437097 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.437168 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.437195 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.437220 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.438064 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.439727 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.459541 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.471203 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.486001 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.492788 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.492886 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d52e5261-12ff-423f-ace5-22da40e004e6-hosts-file\") pod \"node-resolver-sdcbj\" (UID: \"d52e5261-12ff-423f-ace5-22da40e004e6\") " pod="openshift-dns/node-resolver-sdcbj" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.492911 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/43de4639-a922-4182-9377-d1f28b3aa4c1-mcd-auth-proxy-config\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.492934 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz4xs\" (UniqueName: \"kubernetes.io/projected/43de4639-a922-4182-9377-d1f28b3aa4c1-kube-api-access-hz4xs\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.492973 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.492993 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/43de4639-a922-4182-9377-d1f28b3aa4c1-rootfs\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.493015 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.493036 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/43de4639-a922-4182-9377-d1f28b3aa4c1-proxy-tls\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.493061 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddxhv\" (UniqueName: \"kubernetes.io/projected/d52e5261-12ff-423f-ace5-22da40e004e6-kube-api-access-ddxhv\") pod \"node-resolver-sdcbj\" (UID: \"d52e5261-12ff-423f-ace5-22da40e004e6\") " pod="openshift-dns/node-resolver-sdcbj" Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.493181 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:16:56.493147344 +0000 UTC m=+24.359162957 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.493262 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.493303 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:56.493295698 +0000 UTC m=+24.359311321 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.493623 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.493692 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:56.493652568 +0000 UTC m=+24.359668191 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.499545 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.509039 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.523808 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.543135 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.562513 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.578026 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.591286 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.593510 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.593559 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d52e5261-12ff-423f-ace5-22da40e004e6-hosts-file\") pod \"node-resolver-sdcbj\" (UID: \"d52e5261-12ff-423f-ace5-22da40e004e6\") " pod="openshift-dns/node-resolver-sdcbj" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.593592 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/43de4639-a922-4182-9377-d1f28b3aa4c1-mcd-auth-proxy-config\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.593617 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz4xs\" (UniqueName: \"kubernetes.io/projected/43de4639-a922-4182-9377-d1f28b3aa4c1-kube-api-access-hz4xs\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.593637 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.593666 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/43de4639-a922-4182-9377-d1f28b3aa4c1-rootfs\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.593695 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/43de4639-a922-4182-9377-d1f28b3aa4c1-proxy-tls\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.593714 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddxhv\" (UniqueName: \"kubernetes.io/projected/d52e5261-12ff-423f-ace5-22da40e004e6-kube-api-access-ddxhv\") pod \"node-resolver-sdcbj\" (UID: \"d52e5261-12ff-423f-ace5-22da40e004e6\") " pod="openshift-dns/node-resolver-sdcbj" Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.593731 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.593761 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.593776 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.593826 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:56.593808668 +0000 UTC m=+24.459824291 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.594090 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/43de4639-a922-4182-9377-d1f28b3aa4c1-rootfs\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.594172 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d52e5261-12ff-423f-ace5-22da40e004e6-hosts-file\") pod \"node-resolver-sdcbj\" (UID: \"d52e5261-12ff-423f-ace5-22da40e004e6\") " pod="openshift-dns/node-resolver-sdcbj" Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.594251 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.594282 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.594295 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:54 crc kubenswrapper[4735]: E0215 20:16:54.594358 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-15 20:16:56.594340253 +0000 UTC m=+24.460355876 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.594810 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/43de4639-a922-4182-9377-d1f28b3aa4c1-mcd-auth-proxy-config\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.598989 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/43de4639-a922-4182-9377-d1f28b3aa4c1-proxy-tls\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.607323 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.612225 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz4xs\" (UniqueName: \"kubernetes.io/projected/43de4639-a922-4182-9377-d1f28b3aa4c1-kube-api-access-hz4xs\") pod \"machine-config-daemon-kjps6\" (UID: \"43de4639-a922-4182-9377-d1f28b3aa4c1\") " pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.616294 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddxhv\" (UniqueName: \"kubernetes.io/projected/d52e5261-12ff-423f-ace5-22da40e004e6-kube-api-access-ddxhv\") pod \"node-resolver-sdcbj\" (UID: \"d52e5261-12ff-423f-ace5-22da40e004e6\") " pod="openshift-dns/node-resolver-sdcbj" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.631899 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.648238 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.702250 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.745048 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.747131 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-sdcbj" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.751652 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.821183 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.824016 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 01:24:21.765884273 +0000 UTC Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.869671 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.890097 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.890796 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.895211 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.895848 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.896933 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.897538 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.898185 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.899162 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.899827 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.900900 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.901501 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.902836 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.903596 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.904501 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.906843 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.907424 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.909302 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.909763 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.910382 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.911507 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.913757 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.914421 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.917447 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.918163 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.919167 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.920341 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.921009 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.922050 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.922586 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.923225 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.928390 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.928921 4735 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.929096 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.931293 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.931891 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.932386 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.935056 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.935749 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.936695 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.937406 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.939048 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.939551 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.940196 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.945495 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.946555 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.947230 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.948198 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.948963 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.949735 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.950720 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.951280 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.951792 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.956264 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.956895 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.957974 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.958459 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-fds4z"] Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.959076 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-wfjmx"] Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.959309 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x9xmf"] Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.959559 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.959905 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-wfjmx" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.960982 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.970433 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.970633 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.970755 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.970869 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.971043 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.971585 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.975535 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.975592 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.976040 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.979025 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.980414 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.980794 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.981030 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.981254 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 15 20:16:54 crc kubenswrapper[4735]: I0215 20:16:54.981426 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.019024 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.053994 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-15 20:11:54 +0000 UTC, rotation deadline is 2027-01-03 05:27:44.338543155 +0000 UTC Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.054052 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7713h10m49.284493517s for next certificate rotation Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.054400 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.069141 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-sdcbj" event={"ID":"d52e5261-12ff-423f-ace5-22da40e004e6","Type":"ContainerStarted","Data":"9b96c2b039466c79fb17c0ca3fb024e16905d18a4f45d2af71b2215caf7ca63e"} Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.070859 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28"} Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.070888 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"e317f70927b909bba1f4c8cb5b3d4973728315f84ef79b3da113ea5a171a5eb9"} Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.084277 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096651 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-daemon-config\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096684 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmcnx\" (UniqueName: \"kubernetes.io/projected/a20277cf-39a2-453f-ac6d-fb28346b3358-kube-api-access-mmcnx\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096705 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-netns\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096724 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk9bx\" (UniqueName: \"kubernetes.io/projected/755c2f64-a91b-443f-96ca-88c8cab0d656-kube-api-access-nk9bx\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096743 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-system-cni-dir\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096760 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-os-release\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096778 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-os-release\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096794 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-var-lib-kubelet\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096813 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-script-lib\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096829 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5c684f83-2a4f-465d-95f1-43e54317d164-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096847 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpdwr\" (UniqueName: \"kubernetes.io/projected/5c684f83-2a4f-465d-95f1-43e54317d164-kube-api-access-cpdwr\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096864 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a20277cf-39a2-453f-ac6d-fb28346b3358-cni-binary-copy\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096880 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-var-lib-cni-bin\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096898 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-cnibin\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096914 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-netd\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096930 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096971 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/755c2f64-a91b-443f-96ca-88c8cab0d656-ovn-node-metrics-cert\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.096989 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-var-lib-openvswitch\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097004 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-env-overrides\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097041 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-etc-kubernetes\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097057 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-kubelet\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097074 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-var-lib-cni-multus\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097091 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-slash\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097109 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-cni-dir\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097125 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-conf-dir\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097140 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-systemd\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097169 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-ovn-kubernetes\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097195 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097212 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-openvswitch\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097228 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-config\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097250 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-run-netns\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097267 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-etc-openvswitch\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097288 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-node-log\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097314 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5c684f83-2a4f-465d-95f1-43e54317d164-cni-binary-copy\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097332 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-run-multus-certs\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097350 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-log-socket\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097366 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-hostroot\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097390 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-cnibin\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097405 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-system-cni-dir\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097425 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-socket-dir-parent\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097443 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-run-k8s-cni-cncf-io\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097460 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-systemd-units\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097476 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-ovn\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.097493 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-bin\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.104925 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.123166 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.139763 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.152929 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.169802 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.182447 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.193340 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.198845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-cnibin\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200029 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-netd\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200152 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200253 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/755c2f64-a91b-443f-96ca-88c8cab0d656-ovn-node-metrics-cert\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200358 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-var-lib-openvswitch\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200449 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-env-overrides\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200547 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-var-lib-cni-multus\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200641 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-etc-kubernetes\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200727 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-kubelet\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200827 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-slash\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200928 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201043 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-cni-dir\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201147 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-conf-dir\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201244 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-systemd\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201334 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-ovn-kubernetes\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200194 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.200167 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-netd\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201387 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-var-lib-cni-multus\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.199070 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-cnibin\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201407 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-etc-kubernetes\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201530 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-var-lib-openvswitch\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201654 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-systemd\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201692 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-conf-dir\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201726 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-ovn-kubernetes\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201719 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-cni-dir\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201772 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-slash\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.201355 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-kubelet\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202065 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-config\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202142 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-openvswitch\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202210 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-run-netns\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202278 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-etc-openvswitch\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202346 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-node-log\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202544 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5c684f83-2a4f-465d-95f1-43e54317d164-cni-binary-copy\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202645 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-run-multus-certs\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202731 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-log-socket\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202859 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-system-cni-dir\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202935 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-socket-dir-parent\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203032 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-hostroot\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203111 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-cnibin\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203178 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-run-k8s-cni-cncf-io\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203258 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-systemd-units\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203320 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-system-cni-dir\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202480 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-node-log\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202837 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-log-socket\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203223 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5c684f83-2a4f-465d-95f1-43e54317d164-cni-binary-copy\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202098 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-env-overrides\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202506 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-openvswitch\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202554 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-etc-openvswitch\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203466 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-socket-dir-parent\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202657 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203505 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-hostroot\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202591 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-config\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203540 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-cnibin\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202455 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-run-netns\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203572 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-run-k8s-cni-cncf-io\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203594 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-systemd-units\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.202689 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-run-multus-certs\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203695 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-ovn\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203332 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-ovn\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203908 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-bin\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.203996 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-os-release\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204071 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-os-release\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204154 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-daemon-config\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204219 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmcnx\" (UniqueName: \"kubernetes.io/projected/a20277cf-39a2-453f-ac6d-fb28346b3358-kube-api-access-mmcnx\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204289 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-netns\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204357 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk9bx\" (UniqueName: \"kubernetes.io/projected/755c2f64-a91b-443f-96ca-88c8cab0d656-kube-api-access-nk9bx\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204425 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-system-cni-dir\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204490 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a20277cf-39a2-453f-ac6d-fb28346b3358-cni-binary-copy\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204549 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-var-lib-cni-bin\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204609 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-var-lib-kubelet\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204673 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-script-lib\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204744 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5c684f83-2a4f-465d-95f1-43e54317d164-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.204812 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpdwr\" (UniqueName: \"kubernetes.io/projected/5c684f83-2a4f-465d-95f1-43e54317d164-kube-api-access-cpdwr\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.205043 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-bin\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.205174 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-netns\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.205231 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-os-release\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.205272 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-os-release\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.205475 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5c684f83-2a4f-465d-95f1-43e54317d164-system-cni-dir\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.205729 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a20277cf-39a2-453f-ac6d-fb28346b3358-multus-daemon-config\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.205773 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-var-lib-kubelet\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.205798 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a20277cf-39a2-453f-ac6d-fb28346b3358-host-var-lib-cni-bin\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.206243 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-script-lib\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.206557 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5c684f83-2a4f-465d-95f1-43e54317d164-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.207048 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a20277cf-39a2-453f-ac6d-fb28346b3358-cni-binary-copy\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.210053 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.210393 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/755c2f64-a91b-443f-96ca-88c8cab0d656-ovn-node-metrics-cert\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.226569 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmcnx\" (UniqueName: \"kubernetes.io/projected/a20277cf-39a2-453f-ac6d-fb28346b3358-kube-api-access-mmcnx\") pod \"multus-wfjmx\" (UID: \"a20277cf-39a2-453f-ac6d-fb28346b3358\") " pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.226707 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.235297 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpdwr\" (UniqueName: \"kubernetes.io/projected/5c684f83-2a4f-465d-95f1-43e54317d164-kube-api-access-cpdwr\") pod \"multus-additional-cni-plugins-fds4z\" (UID: \"5c684f83-2a4f-465d-95f1-43e54317d164\") " pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.235302 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk9bx\" (UniqueName: \"kubernetes.io/projected/755c2f64-a91b-443f-96ca-88c8cab0d656-kube-api-access-nk9bx\") pod \"ovnkube-node-x9xmf\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.247108 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.266675 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.302898 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fds4z" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.311577 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-wfjmx" Feb 15 20:16:55 crc kubenswrapper[4735]: W0215 20:16:55.316474 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c684f83_2a4f_465d_95f1_43e54317d164.slice/crio-1438bc79b4b092f0ff7272a8f88300aa7867c25b8fdec6ffa8f3e9a1b0e4a212 WatchSource:0}: Error finding container 1438bc79b4b092f0ff7272a8f88300aa7867c25b8fdec6ffa8f3e9a1b0e4a212: Status 404 returned error can't find the container with id 1438bc79b4b092f0ff7272a8f88300aa7867c25b8fdec6ffa8f3e9a1b0e4a212 Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.318294 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.824776 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 03:45:27.460534117 +0000 UTC Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.886802 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.886857 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:16:55 crc kubenswrapper[4735]: I0215 20:16:55.886895 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:16:55 crc kubenswrapper[4735]: E0215 20:16:55.886996 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:16:55 crc kubenswrapper[4735]: E0215 20:16:55.887092 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:16:55 crc kubenswrapper[4735]: E0215 20:16:55.887130 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.075390 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-sdcbj" event={"ID":"d52e5261-12ff-423f-ace5-22da40e004e6","Type":"ContainerStarted","Data":"6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e"} Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.077548 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c684f83-2a4f-465d-95f1-43e54317d164" containerID="84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714" exitCode=0 Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.077624 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" event={"ID":"5c684f83-2a4f-465d-95f1-43e54317d164","Type":"ContainerDied","Data":"84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714"} Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.077658 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" event={"ID":"5c684f83-2a4f-465d-95f1-43e54317d164","Type":"ContainerStarted","Data":"1438bc79b4b092f0ff7272a8f88300aa7867c25b8fdec6ffa8f3e9a1b0e4a212"} Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.079265 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48" exitCode=0 Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.079356 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48"} Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.079421 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"2d31391d2ab043f7a2e016ed1ae052f7e5beead951464715e04f361c04e09a1d"} Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.096432 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wfjmx" event={"ID":"a20277cf-39a2-453f-ac6d-fb28346b3358","Type":"ContainerStarted","Data":"4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9"} Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.096488 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wfjmx" event={"ID":"a20277cf-39a2-453f-ac6d-fb28346b3358","Type":"ContainerStarted","Data":"967e348bbae51111b50d9b6a923ab0e19b4bfeb7083eaa58d81cb2fe1787a2b0"} Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.105140 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6"} Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.109646 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0"} Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.124010 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.150604 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.224336 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.259333 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.305134 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.345351 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.379117 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.393157 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.408258 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.422454 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.436893 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.456286 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.486386 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.517648 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.517816 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:17:00.517786253 +0000 UTC m=+28.383801876 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.517974 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.518077 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.518131 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.518239 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.518017 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.518263 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:00.518247786 +0000 UTC m=+28.384263409 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.518452 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:00.518441881 +0000 UTC m=+28.384457504 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.534734 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.544783 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-9ls74"] Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.545195 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.547212 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.547236 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.548905 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.549214 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.549956 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.561651 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.575431 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.593772 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.604940 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.615909 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.618610 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.618667 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1df5b8f-f6f3-45a6-99d2-088da4f8981b-host\") pod \"node-ca-9ls74\" (UID: \"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\") " pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.618690 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d1df5b8f-f6f3-45a6-99d2-088da4f8981b-serviceca\") pod \"node-ca-9ls74\" (UID: \"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\") " pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.618711 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pvb8\" (UniqueName: \"kubernetes.io/projected/d1df5b8f-f6f3-45a6-99d2-088da4f8981b-kube-api-access-6pvb8\") pod \"node-ca-9ls74\" (UID: \"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\") " pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.618758 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.618876 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.618901 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.618915 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.618994 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:00.618976852 +0000 UTC m=+28.484992465 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.619065 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.619286 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.619298 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:56 crc kubenswrapper[4735]: E0215 20:16:56.619374 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:00.619356102 +0000 UTC m=+28.485371715 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.630536 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.648856 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.665215 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.679175 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.693368 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.704429 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.718634 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.720157 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1df5b8f-f6f3-45a6-99d2-088da4f8981b-host\") pod \"node-ca-9ls74\" (UID: \"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\") " pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.720202 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d1df5b8f-f6f3-45a6-99d2-088da4f8981b-serviceca\") pod \"node-ca-9ls74\" (UID: \"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\") " pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.720223 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pvb8\" (UniqueName: \"kubernetes.io/projected/d1df5b8f-f6f3-45a6-99d2-088da4f8981b-kube-api-access-6pvb8\") pod \"node-ca-9ls74\" (UID: \"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\") " pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.720324 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1df5b8f-f6f3-45a6-99d2-088da4f8981b-host\") pod \"node-ca-9ls74\" (UID: \"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\") " pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.721409 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d1df5b8f-f6f3-45a6-99d2-088da4f8981b-serviceca\") pod \"node-ca-9ls74\" (UID: \"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\") " pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.729151 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.739601 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pvb8\" (UniqueName: \"kubernetes.io/projected/d1df5b8f-f6f3-45a6-99d2-088da4f8981b-kube-api-access-6pvb8\") pod \"node-ca-9ls74\" (UID: \"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\") " pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.742395 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.759568 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.773530 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.783573 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.796307 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.811478 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.825036 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 11:42:59.522351303 +0000 UTC Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.828305 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.851871 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.867158 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.883900 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.899308 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:56Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:56 crc kubenswrapper[4735]: I0215 20:16:56.965747 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-9ls74" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.122286 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c684f83-2a4f-465d-95f1-43e54317d164" containerID="b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d" exitCode=0 Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.122349 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" event={"ID":"5c684f83-2a4f-465d-95f1-43e54317d164","Type":"ContainerDied","Data":"b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d"} Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.127773 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-9ls74" event={"ID":"d1df5b8f-f6f3-45a6-99d2-088da4f8981b","Type":"ContainerStarted","Data":"74acc929679e7c61e21057aa1c9532e33399e2e2a4732cc1d43bd86fd99d45fa"} Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.136171 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e"} Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.136221 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc"} Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.136235 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62"} Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.139054 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.152554 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.170114 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.189242 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.212308 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.224642 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.238939 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.253312 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.268082 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.299994 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.336435 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.354877 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.368442 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.387791 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:57Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.826108 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 04:01:37.136551058 +0000 UTC Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.886007 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.886104 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:16:57 crc kubenswrapper[4735]: E0215 20:16:57.886211 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:16:57 crc kubenswrapper[4735]: E0215 20:16:57.886316 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:16:57 crc kubenswrapper[4735]: I0215 20:16:57.886644 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:16:57 crc kubenswrapper[4735]: E0215 20:16:57.886909 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.143351 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920"} Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.143401 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0"} Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.143415 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55"} Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.145800 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c684f83-2a4f-465d-95f1-43e54317d164" containerID="5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa" exitCode=0 Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.145874 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" event={"ID":"5c684f83-2a4f-465d-95f1-43e54317d164","Type":"ContainerDied","Data":"5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa"} Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.149473 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-9ls74" event={"ID":"d1df5b8f-f6f3-45a6-99d2-088da4f8981b","Type":"ContainerStarted","Data":"3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c"} Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.175908 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.194131 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.214097 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.231020 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.244060 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.256090 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.270745 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.289922 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.304519 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.403457 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.428833 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.448574 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.491624 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.505174 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.526559 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.542651 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.555693 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.568329 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.586891 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.604570 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.628553 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.639336 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.651152 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.663817 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.677344 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.693198 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.706292 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.718324 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:58Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:58 crc kubenswrapper[4735]: I0215 20:16:58.827349 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 02:12:22.27660018 +0000 UTC Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.157693 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c684f83-2a4f-465d-95f1-43e54317d164" containerID="52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb" exitCode=0 Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.157818 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" event={"ID":"5c684f83-2a4f-465d-95f1-43e54317d164","Type":"ContainerDied","Data":"52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.175965 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.193280 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.208228 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.220916 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.223069 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.223109 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.223117 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.223220 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.231866 4735 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.232059 4735 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.233058 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.233120 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.233134 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.233162 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.233177 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.235491 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: E0215 20:16:59.249619 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.251115 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.252703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.252744 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.252755 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.252772 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.252783 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: E0215 20:16:59.264766 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.266491 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.268344 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.268383 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.268394 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.268414 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.268425 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: E0215 20:16:59.279219 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.283426 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.283499 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.283553 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.283576 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.283592 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.283845 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: E0215 20:16:59.296732 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.299302 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.300995 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.301014 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.301021 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.301035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.301045 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.318329 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: E0215 20:16:59.319134 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: E0215 20:16:59.319394 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.324163 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.324230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.324274 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.324302 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.324318 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.332712 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.348830 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.359957 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.372447 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.385423 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:16:59Z is after 2025-08-24T17:21:41Z" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.427979 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.428020 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.428029 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.428045 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.428056 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.530548 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.530587 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.530596 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.530612 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.530623 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.633131 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.633179 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.633188 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.633204 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.633215 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.736050 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.736091 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.736104 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.736121 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.736132 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.828303 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 06:02:31.469684025 +0000 UTC Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.838086 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.838144 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.838162 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.838189 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.838207 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.886003 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.886081 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:16:59 crc kubenswrapper[4735]: E0215 20:16:59.886150 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.886091 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:16:59 crc kubenswrapper[4735]: E0215 20:16:59.886235 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:16:59 crc kubenswrapper[4735]: E0215 20:16:59.886360 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.941331 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.941848 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.941872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.941898 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:16:59 crc kubenswrapper[4735]: I0215 20:16:59.941917 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:16:59Z","lastTransitionTime":"2026-02-15T20:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.045157 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.045198 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.045211 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.045229 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.045240 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.148562 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.148629 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.148646 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.148671 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.148685 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.165790 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.168932 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c684f83-2a4f-465d-95f1-43e54317d164" containerID="625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71" exitCode=0 Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.168988 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" event={"ID":"5c684f83-2a4f-465d-95f1-43e54317d164","Type":"ContainerDied","Data":"625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.199370 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.231591 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.250856 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.250895 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.250906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.250923 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.250934 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.267992 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.283010 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.295612 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.310287 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.324160 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.338990 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.347212 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.353679 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.353733 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.353748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.353768 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.353782 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.359735 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.374599 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.396406 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.413075 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.424535 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:00Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.456598 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.456650 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.456659 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.456683 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.456697 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.560582 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.560645 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.560657 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.560680 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.560692 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.610348 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.610587 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:17:08.610544497 +0000 UTC m=+36.476560160 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.610689 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.610788 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.610990 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.611079 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:08.611063671 +0000 UTC m=+36.477079334 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.611133 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.611240 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:08.611214965 +0000 UTC m=+36.477230598 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.663442 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.663523 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.663542 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.663570 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.663585 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.711727 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.711834 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.712092 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.712123 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.712145 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.712178 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.712259 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.712542 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.712268 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:08.712238629 +0000 UTC m=+36.578254482 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:00 crc kubenswrapper[4735]: E0215 20:17:00.712991 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:08.712922749 +0000 UTC m=+36.578938412 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.766827 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.767263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.767366 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.767462 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.767558 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.829211 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 20:08:18.718448078 +0000 UTC Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.871211 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.871311 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.871332 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.871365 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.871387 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.975543 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.975635 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.975663 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.975698 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:00 crc kubenswrapper[4735]: I0215 20:17:00.975722 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:00Z","lastTransitionTime":"2026-02-15T20:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.080014 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.080083 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.080101 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.080127 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.080145 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:01Z","lastTransitionTime":"2026-02-15T20:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.186000 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.186055 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.186072 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.186098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.186119 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:01Z","lastTransitionTime":"2026-02-15T20:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.191332 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c684f83-2a4f-465d-95f1-43e54317d164" containerID="7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5" exitCode=0 Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.191423 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" event={"ID":"5c684f83-2a4f-465d-95f1-43e54317d164","Type":"ContainerDied","Data":"7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.222320 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.248486 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.277089 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.289769 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.289818 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.289832 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.289853 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.289866 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:01Z","lastTransitionTime":"2026-02-15T20:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.302537 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.319743 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.337072 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.355014 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.370549 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.382287 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.392552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.392597 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.392609 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.392629 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.392641 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:01Z","lastTransitionTime":"2026-02-15T20:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.396464 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.416219 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.442011 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.463692 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.482269 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:01Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.495202 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.495270 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.495285 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.495312 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.495326 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:01Z","lastTransitionTime":"2026-02-15T20:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.599215 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.599277 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.599296 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.599327 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.599347 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:01Z","lastTransitionTime":"2026-02-15T20:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.703023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.703092 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.703112 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.703146 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.703165 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:01Z","lastTransitionTime":"2026-02-15T20:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.807604 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.807694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.807717 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.807746 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.807772 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:01Z","lastTransitionTime":"2026-02-15T20:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.830347 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 02:03:38.256273267 +0000 UTC Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.886529 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.886562 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.886724 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:01 crc kubenswrapper[4735]: E0215 20:17:01.886921 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:01 crc kubenswrapper[4735]: E0215 20:17:01.887751 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:01 crc kubenswrapper[4735]: E0215 20:17:01.887863 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.912472 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.912541 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.912564 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.912591 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:01 crc kubenswrapper[4735]: I0215 20:17:01.912609 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:01Z","lastTransitionTime":"2026-02-15T20:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.016712 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.017369 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.017396 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.017454 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.017484 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.120528 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.120591 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.120610 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.120637 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.120657 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.203282 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" event={"ID":"5c684f83-2a4f-465d-95f1-43e54317d164","Type":"ContainerStarted","Data":"dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.223169 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.223781 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.224093 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.224576 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.224623 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.224641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.224662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.224680 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.240911 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.259592 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.270252 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.272310 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.282303 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.303400 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.321903 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.326723 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.326750 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.326765 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.326786 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.326800 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.343502 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.365533 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.386903 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.405081 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.422720 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.429090 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.429119 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.429132 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.429151 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.429162 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.444532 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.460147 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.468836 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.476791 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.497481 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.513670 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.526826 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.531120 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.531165 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.531181 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.531204 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.531219 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.540173 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.557546 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.576691 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.596449 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.615650 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.634085 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.634138 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.634152 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.634179 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.634195 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.639133 4735 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.640529 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ovn-kubernetes/pods/ovnkube-node-x9xmf/status\": read tcp 38.102.83.82:35416->38.102.83.82:6443: use of closed network connection" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.697132 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.713666 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.738704 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.741171 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.741217 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.741230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.741250 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.741265 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.755903 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.771282 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.787045 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.830866 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 05:39:48.361140587 +0000 UTC Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.845361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.845423 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.845439 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.845464 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.845480 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.908382 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.933535 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.949438 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.949496 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.949518 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.949549 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.949570 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:02Z","lastTransitionTime":"2026-02-15T20:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.958447 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:02 crc kubenswrapper[4735]: I0215 20:17:02.989689 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:02Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.023837 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.050411 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.052391 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.052487 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.052513 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.052579 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.052602 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.076712 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.098084 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.125779 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.143930 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.156547 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.156624 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.156644 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.156673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.156692 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.163335 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.185723 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.210714 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.227116 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.239930 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:03Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.260227 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.260275 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.260288 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.260310 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.260324 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.364333 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.364713 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.364886 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.365083 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.365229 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.469229 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.469592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.469736 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.470096 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.470242 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.573440 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.573510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.573529 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.573557 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.573577 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.677844 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.677918 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.677935 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.677993 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.678011 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.780882 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.780939 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.780988 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.781020 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.781038 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.831551 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 21:45:04.036670751 +0000 UTC Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.886170 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:03 crc kubenswrapper[4735]: E0215 20:17:03.886429 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.886495 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.886529 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.886630 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:03 crc kubenswrapper[4735]: E0215 20:17:03.886728 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.886545 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: E0215 20:17:03.886863 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.886913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.886971 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.886993 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.994674 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.994773 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.994803 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.994837 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:03 crc kubenswrapper[4735]: I0215 20:17:03.994863 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:03Z","lastTransitionTime":"2026-02-15T20:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.098719 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.098778 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.098795 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.098822 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.098842 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:04Z","lastTransitionTime":"2026-02-15T20:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.202760 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.202823 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.202841 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.202871 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.202892 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:04Z","lastTransitionTime":"2026-02-15T20:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.231468 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.307126 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.307228 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.307250 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.307281 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.307304 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:04Z","lastTransitionTime":"2026-02-15T20:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.410898 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.410992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.411023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.411055 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.411079 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:04Z","lastTransitionTime":"2026-02-15T20:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.513691 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.513736 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.513748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.513767 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.513783 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:04Z","lastTransitionTime":"2026-02-15T20:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.616321 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.616372 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.616382 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.616399 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.616409 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:04Z","lastTransitionTime":"2026-02-15T20:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.718872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.718955 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.718967 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.718992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.719002 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:04Z","lastTransitionTime":"2026-02-15T20:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.821446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.821487 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.821500 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.821516 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.821526 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:04Z","lastTransitionTime":"2026-02-15T20:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.831791 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 09:45:35.390829166 +0000 UTC Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.924646 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.924709 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.924733 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.924882 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:04 crc kubenswrapper[4735]: I0215 20:17:04.924936 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:04Z","lastTransitionTime":"2026-02-15T20:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.027239 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.027327 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.027349 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.027431 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.027450 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.129821 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.129864 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.129877 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.129895 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.129906 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.233580 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.233662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.233684 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.233717 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.233748 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.336622 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.337242 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.337263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.337288 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.337306 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.441086 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.441149 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.441169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.441196 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.441213 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.543854 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.543911 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.543933 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.544002 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.544028 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.647694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.647758 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.647777 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.647808 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.647826 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.750940 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.751031 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.751048 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.751070 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.751087 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.831896 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 09:28:26.535909185 +0000 UTC Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.854378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.854453 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.854470 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.854495 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.854515 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.886227 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.886353 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:05 crc kubenswrapper[4735]: E0215 20:17:05.886443 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.886491 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:05 crc kubenswrapper[4735]: E0215 20:17:05.886651 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:05 crc kubenswrapper[4735]: E0215 20:17:05.886750 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.957658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.957777 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.957798 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.957865 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:05 crc kubenswrapper[4735]: I0215 20:17:05.957887 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:05Z","lastTransitionTime":"2026-02-15T20:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.060972 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.061045 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.061065 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.061090 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.061113 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.164556 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.164675 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.164756 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.165476 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.165686 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.247802 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/0.log" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.252772 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1" exitCode=1 Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.252838 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.254130 4735 scope.go:117] "RemoveContainer" containerID="cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.269024 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.269143 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.269170 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.269201 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.269224 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.280043 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.302106 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.320612 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.347554 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.370062 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.371995 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.372053 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.372073 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.372104 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.372126 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.388323 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.407300 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.432394 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.470868 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:05Z\\\",\\\"message\\\":\\\"5 20:17:05.412196 5957 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0215 20:17:05.412377 5957 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412698 5957 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412872 5957 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.413376 5957 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:05.413422 5957 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:05.413459 5957 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:05.413469 5957 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:05.413473 5957 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:05.413492 5957 factory.go:656] Stopping watch factory\\\\nI0215 20:17:05.413514 5957 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:05.413528 5957 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:05.413539 5957 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.476214 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.476289 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.476323 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.476368 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.476398 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.499038 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.519228 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.547537 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.576820 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.580712 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.580753 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.580773 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.580799 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.580819 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.604004 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:06Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.685039 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.685119 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.685150 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.685182 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.685206 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.789820 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.789866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.789885 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.789911 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.789930 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.832729 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 05:36:16.472628305 +0000 UTC Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.892936 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.893014 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.893032 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.893057 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.893100 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.998576 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.998637 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.998655 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.998679 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:06 crc kubenswrapper[4735]: I0215 20:17:06.998698 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:06Z","lastTransitionTime":"2026-02-15T20:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.102035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.102104 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.102127 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.102161 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.102184 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:07Z","lastTransitionTime":"2026-02-15T20:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.205550 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.205605 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.205620 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.205641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.205656 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:07Z","lastTransitionTime":"2026-02-15T20:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.259261 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/0.log" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.262170 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.262405 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.275112 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.287624 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.298833 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.308893 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.309015 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.309042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.309070 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.309086 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:07Z","lastTransitionTime":"2026-02-15T20:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.310515 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.322892 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.336100 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.351415 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.372818 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:05Z\\\",\\\"message\\\":\\\"5 20:17:05.412196 5957 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0215 20:17:05.412377 5957 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412698 5957 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412872 5957 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.413376 5957 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:05.413422 5957 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:05.413459 5957 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:05.413469 5957 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:05.413473 5957 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:05.413492 5957 factory.go:656] Stopping watch factory\\\\nI0215 20:17:05.413514 5957 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:05.413528 5957 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:05.413539 5957 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.390350 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.407687 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.411351 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.411415 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.411429 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.411660 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.411673 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:07Z","lastTransitionTime":"2026-02-15T20:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.425636 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.440626 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.457361 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.472630 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:07Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.513639 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.513678 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.513688 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.513704 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.513716 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:07Z","lastTransitionTime":"2026-02-15T20:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.615723 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.615780 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.615795 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.615817 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.615833 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:07Z","lastTransitionTime":"2026-02-15T20:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.718347 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.718418 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.718438 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.718473 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.718498 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:07Z","lastTransitionTime":"2026-02-15T20:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.830760 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.830842 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.830867 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.830899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.830919 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:07Z","lastTransitionTime":"2026-02-15T20:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.833005 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 02:32:44.554042823 +0000 UTC Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.886827 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.886881 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.887036 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:07 crc kubenswrapper[4735]: E0215 20:17:07.887115 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:07 crc kubenswrapper[4735]: E0215 20:17:07.887284 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:07 crc kubenswrapper[4735]: E0215 20:17:07.887451 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.934278 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.934328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.934343 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.934361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:07 crc kubenswrapper[4735]: I0215 20:17:07.934375 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:07Z","lastTransitionTime":"2026-02-15T20:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.037321 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.037377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.037399 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.037425 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.037446 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.140913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.141100 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.141121 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.141149 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.141170 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.244483 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.244560 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.244580 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.244609 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.244630 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.273709 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/1.log" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.277001 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/0.log" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.283929 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2" exitCode=1 Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.284026 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.284163 4735 scope.go:117] "RemoveContainer" containerID="cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.285779 4735 scope.go:117] "RemoveContainer" containerID="2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2" Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.286114 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.308140 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.331350 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.347806 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.347874 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.347892 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.347936 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.347978 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.354862 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.371890 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.390379 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.412915 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.434634 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.450776 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.451384 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.451458 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.451482 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.451512 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.451532 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.469315 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.490131 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.521263 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:05Z\\\",\\\"message\\\":\\\"5 20:17:05.412196 5957 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0215 20:17:05.412377 5957 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412698 5957 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412872 5957 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.413376 5957 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:05.413422 5957 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:05.413459 5957 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:05.413469 5957 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:05.413473 5957 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:05.413492 5957 factory.go:656] Stopping watch factory\\\\nI0215 20:17:05.413514 5957 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:05.413528 5957 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:05.413539 5957 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:07Z\\\",\\\"message\\\":\\\" 6094 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0215 20:17:07.579544 6094 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:07.579626 6094 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0215 20:17:07.579723 6094 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:07.579765 6094 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0215 20:17:07.579782 6094 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0215 20:17:07.579812 6094 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:07.579822 6094 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:07.579833 6094 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:07.579789 6094 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:07.579835 6094 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0215 20:17:07.579863 6094 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:07.579886 6094 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0215 20:17:07.579934 6094 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0215 20:17:07.579974 6094 factory.go:656] Stopping watch factory\\\\nI0215 20:17:07.580011 6094 ovnkube.go:599] Stopped ovnkube\\\\nI02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.539432 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.550392 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm"] Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.553108 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.555528 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.556434 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.557117 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.557214 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.557241 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.557273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.557314 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.563317 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.591932 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.606908 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/92a41601-6d64-46c3-a9ea-2fe0427d884b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.607033 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vbqz\" (UniqueName: \"kubernetes.io/projected/92a41601-6d64-46c3-a9ea-2fe0427d884b-kube-api-access-8vbqz\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.607201 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/92a41601-6d64-46c3-a9ea-2fe0427d884b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.607302 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/92a41601-6d64-46c3-a9ea-2fe0427d884b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.608781 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.634211 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.661252 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.661306 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.661326 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.661352 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.661372 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.665153 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:05Z\\\",\\\"message\\\":\\\"5 20:17:05.412196 5957 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0215 20:17:05.412377 5957 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412698 5957 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412872 5957 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.413376 5957 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:05.413422 5957 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:05.413459 5957 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:05.413469 5957 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:05.413473 5957 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:05.413492 5957 factory.go:656] Stopping watch factory\\\\nI0215 20:17:05.413514 5957 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:05.413528 5957 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:05.413539 5957 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:07Z\\\",\\\"message\\\":\\\" 6094 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0215 20:17:07.579544 6094 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:07.579626 6094 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0215 20:17:07.579723 6094 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:07.579765 6094 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0215 20:17:07.579782 6094 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0215 20:17:07.579812 6094 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:07.579822 6094 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:07.579833 6094 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:07.579789 6094 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:07.579835 6094 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0215 20:17:07.579863 6094 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:07.579886 6094 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0215 20:17:07.579934 6094 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0215 20:17:07.579974 6094 factory.go:656] Stopping watch factory\\\\nI0215 20:17:07.580011 6094 ovnkube.go:599] Stopped ovnkube\\\\nI02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.687267 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.706675 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.707992 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.708190 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.708274 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.708317 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/92a41601-6d64-46c3-a9ea-2fe0427d884b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.708362 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/92a41601-6d64-46c3-a9ea-2fe0427d884b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.708417 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/92a41601-6d64-46c3-a9ea-2fe0427d884b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.708479 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vbqz\" (UniqueName: \"kubernetes.io/projected/92a41601-6d64-46c3-a9ea-2fe0427d884b-kube-api-access-8vbqz\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.708820 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.709112 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.708935 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:17:24.708888929 +0000 UTC m=+52.574904592 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.709298 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:24.709257349 +0000 UTC m=+52.575273152 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.709347 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:24.709316691 +0000 UTC m=+52.575332564 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.710128 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/92a41601-6d64-46c3-a9ea-2fe0427d884b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.712593 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/92a41601-6d64-46c3-a9ea-2fe0427d884b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.717345 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/92a41601-6d64-46c3-a9ea-2fe0427d884b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.732544 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.748809 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vbqz\" (UniqueName: \"kubernetes.io/projected/92a41601-6d64-46c3-a9ea-2fe0427d884b-kube-api-access-8vbqz\") pod \"ovnkube-control-plane-749d76644c-q6fbm\" (UID: \"92a41601-6d64-46c3-a9ea-2fe0427d884b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.757763 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.765243 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.765568 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.765717 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.765854 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.766026 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.777292 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.797432 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.810092 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.810187 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.810367 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.810411 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.810451 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.810487 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.810417 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.810556 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.810568 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:24.810541011 +0000 UTC m=+52.676556674 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:08 crc kubenswrapper[4735]: E0215 20:17:08.810630 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:24.810605802 +0000 UTC m=+52.676621445 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.813587 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.833699 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 12:27:39.361503357 +0000 UTC Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.835337 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.856624 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.868773 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.868855 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.868875 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.868902 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.868919 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.875066 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.880342 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.907901 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.933302 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:08Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.972668 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.972739 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.972762 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.972791 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:08 crc kubenswrapper[4735]: I0215 20:17:08.972812 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:08Z","lastTransitionTime":"2026-02-15T20:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.076133 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.076189 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.076206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.076230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.076248 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.182224 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.182653 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.182670 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.182694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.182714 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.286975 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.287011 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.287020 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.287036 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.287051 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.288387 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" event={"ID":"92a41601-6d64-46c3-a9ea-2fe0427d884b","Type":"ContainerStarted","Data":"bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.288430 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" event={"ID":"92a41601-6d64-46c3-a9ea-2fe0427d884b","Type":"ContainerStarted","Data":"9c0344ad29b6ffaa04cb9b31872db3a8593dca47b56163a9fbb1a5425b16054c"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.290003 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/1.log" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.391778 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.391869 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.391894 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.391925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.391977 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.495441 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.495488 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.495505 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.495530 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.495548 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.519443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.519475 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.519489 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.519507 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.519521 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.538534 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.547723 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.547758 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.547773 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.547791 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.547804 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.571843 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.576117 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.576155 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.576168 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.576185 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.576196 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.596026 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.600508 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.600562 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.600572 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.600588 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.600599 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.615501 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.620156 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.620197 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.620206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.620222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.620231 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.632586 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.632797 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.634863 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.634930 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.634990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.635016 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.635034 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.689645 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7vfs4"] Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.690141 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.690206 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.716240 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.720701 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.720768 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgvvs\" (UniqueName: \"kubernetes.io/projected/96681837-f2a3-42d2-9653-a3997f396291-kube-api-access-zgvvs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.738544 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.738648 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.738749 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.738812 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.738842 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.741834 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.757176 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.778663 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.800285 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.821958 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgvvs\" (UniqueName: \"kubernetes.io/projected/96681837-f2a3-42d2-9653-a3997f396291-kube-api-access-zgvvs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.822043 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.822194 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.822276 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs podName:96681837-f2a3-42d2-9653-a3997f396291 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:10.322253414 +0000 UTC m=+38.188269047 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs") pod "network-metrics-daemon-7vfs4" (UID: "96681837-f2a3-42d2-9653-a3997f396291") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.834630 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 02:08:30.432717111 +0000 UTC Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.840269 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:05Z\\\",\\\"message\\\":\\\"5 20:17:05.412196 5957 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0215 20:17:05.412377 5957 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412698 5957 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412872 5957 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.413376 5957 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:05.413422 5957 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:05.413459 5957 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:05.413469 5957 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:05.413473 5957 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:05.413492 5957 factory.go:656] Stopping watch factory\\\\nI0215 20:17:05.413514 5957 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:05.413528 5957 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:05.413539 5957 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:07Z\\\",\\\"message\\\":\\\" 6094 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0215 20:17:07.579544 6094 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:07.579626 6094 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0215 20:17:07.579723 6094 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:07.579765 6094 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0215 20:17:07.579782 6094 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0215 20:17:07.579812 6094 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:07.579822 6094 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:07.579833 6094 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:07.579789 6094 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:07.579835 6094 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0215 20:17:07.579863 6094 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:07.579886 6094 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0215 20:17:07.579934 6094 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0215 20:17:07.579974 6094 factory.go:656] Stopping watch factory\\\\nI0215 20:17:07.580011 6094 ovnkube.go:599] Stopped ovnkube\\\\nI02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.841813 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.841840 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.841851 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.841870 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.841882 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.859306 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgvvs\" (UniqueName: \"kubernetes.io/projected/96681837-f2a3-42d2-9653-a3997f396291-kube-api-access-zgvvs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.860929 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.877473 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.886907 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.887011 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.886921 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.887199 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.887363 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:09 crc kubenswrapper[4735]: E0215 20:17:09.887664 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.898598 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.922157 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.940722 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.944696 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.944734 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.944746 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.944769 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.944783 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:09Z","lastTransitionTime":"2026-02-15T20:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.961373 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.980229 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:09 crc kubenswrapper[4735]: I0215 20:17:09.997878 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:09Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.019629 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.038760 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.047884 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.047985 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.048005 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.048034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.048058 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.151634 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.151706 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.151726 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.151752 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.151772 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.255271 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.255338 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.255357 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.255386 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.255404 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.299616 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" event={"ID":"92a41601-6d64-46c3-a9ea-2fe0427d884b","Type":"ContainerStarted","Data":"755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.323751 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.327357 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:10 crc kubenswrapper[4735]: E0215 20:17:10.327634 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:10 crc kubenswrapper[4735]: E0215 20:17:10.327767 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs podName:96681837-f2a3-42d2-9653-a3997f396291 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:11.327733306 +0000 UTC m=+39.193748969 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs") pod "network-metrics-daemon-7vfs4" (UID: "96681837-f2a3-42d2-9653-a3997f396291") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.345611 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.358883 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.359005 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.359028 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.359061 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.359085 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.364001 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.384728 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.406697 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.427063 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.450039 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.463640 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.463702 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.463721 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.463751 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.463771 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.489565 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:05Z\\\",\\\"message\\\":\\\"5 20:17:05.412196 5957 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0215 20:17:05.412377 5957 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412698 5957 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412872 5957 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.413376 5957 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:05.413422 5957 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:05.413459 5957 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:05.413469 5957 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:05.413473 5957 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:05.413492 5957 factory.go:656] Stopping watch factory\\\\nI0215 20:17:05.413514 5957 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:05.413528 5957 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:05.413539 5957 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:07Z\\\",\\\"message\\\":\\\" 6094 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0215 20:17:07.579544 6094 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:07.579626 6094 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0215 20:17:07.579723 6094 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:07.579765 6094 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0215 20:17:07.579782 6094 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0215 20:17:07.579812 6094 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:07.579822 6094 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:07.579833 6094 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:07.579789 6094 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:07.579835 6094 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0215 20:17:07.579863 6094 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:07.579886 6094 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0215 20:17:07.579934 6094 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0215 20:17:07.579974 6094 factory.go:656] Stopping watch factory\\\\nI0215 20:17:07.580011 6094 ovnkube.go:599] Stopped ovnkube\\\\nI02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.509123 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.536004 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.557925 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.566638 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.566825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.566986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.567166 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.567337 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.587228 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.609584 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.632561 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.650113 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.667764 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:10Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.670919 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.671239 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.671847 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.672177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.672380 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.775343 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.775405 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.775425 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.775452 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.775473 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.835182 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 13:09:29.178757417 +0000 UTC Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.879246 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.879308 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.879329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.879356 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.879377 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.982299 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.982370 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.982387 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.982412 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:10 crc kubenswrapper[4735]: I0215 20:17:10.982433 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:10Z","lastTransitionTime":"2026-02-15T20:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.085483 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.085608 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.085628 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.085652 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.085675 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:11Z","lastTransitionTime":"2026-02-15T20:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.189292 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.189564 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.189694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.189815 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.189931 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:11Z","lastTransitionTime":"2026-02-15T20:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.293301 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.293367 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.293384 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.293410 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.293433 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:11Z","lastTransitionTime":"2026-02-15T20:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.340379 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:11 crc kubenswrapper[4735]: E0215 20:17:11.340746 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:11 crc kubenswrapper[4735]: E0215 20:17:11.340859 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs podName:96681837-f2a3-42d2-9653-a3997f396291 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:13.340825156 +0000 UTC m=+41.206840809 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs") pod "network-metrics-daemon-7vfs4" (UID: "96681837-f2a3-42d2-9653-a3997f396291") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.397746 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.397801 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.397820 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.397849 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.397866 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:11Z","lastTransitionTime":"2026-02-15T20:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.501939 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.502062 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.502082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.502114 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.502134 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:11Z","lastTransitionTime":"2026-02-15T20:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.605317 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.605417 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.605437 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.605462 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.605479 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:11Z","lastTransitionTime":"2026-02-15T20:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.708842 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.708910 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.708928 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.708983 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.709003 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:11Z","lastTransitionTime":"2026-02-15T20:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.812463 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.812543 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.812568 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.812602 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.812625 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:11Z","lastTransitionTime":"2026-02-15T20:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.835929 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 04:20:14.606127263 +0000 UTC Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.886168 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.886303 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.886331 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.886391 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:11 crc kubenswrapper[4735]: E0215 20:17:11.886503 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:11 crc kubenswrapper[4735]: E0215 20:17:11.886589 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:11 crc kubenswrapper[4735]: E0215 20:17:11.886939 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:11 crc kubenswrapper[4735]: E0215 20:17:11.886802 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.915762 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.915827 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.915845 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.915871 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:11 crc kubenswrapper[4735]: I0215 20:17:11.915890 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:11Z","lastTransitionTime":"2026-02-15T20:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.019349 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.019405 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.019422 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.019447 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.019466 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.122781 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.122848 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.122866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.122892 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.122915 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.226115 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.226196 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.226216 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.226246 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.226265 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.328464 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.328531 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.328552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.328581 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.328600 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.432441 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.432509 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.432531 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.432560 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.432579 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.535352 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.535393 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.535402 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.535417 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.535427 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.639046 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.639121 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.639138 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.639167 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.639185 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.742665 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.742750 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.742771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.742802 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.742823 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.837647 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 03:37:34.903751229 +0000 UTC Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.846888 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.846994 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.847014 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.847040 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.847060 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.910151 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:12Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.931771 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:12Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.950623 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.950676 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.950695 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.950721 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.950742 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:12Z","lastTransitionTime":"2026-02-15T20:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.951695 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:12Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.972411 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:12Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:12 crc kubenswrapper[4735]: I0215 20:17:12.993622 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:12Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.015515 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.039619 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.053242 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.053301 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.053317 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.053349 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.053374 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.070328 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:05Z\\\",\\\"message\\\":\\\"5 20:17:05.412196 5957 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0215 20:17:05.412377 5957 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412698 5957 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412872 5957 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.413376 5957 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:05.413422 5957 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:05.413459 5957 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:05.413469 5957 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:05.413473 5957 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:05.413492 5957 factory.go:656] Stopping watch factory\\\\nI0215 20:17:05.413514 5957 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:05.413528 5957 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:05.413539 5957 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:07Z\\\",\\\"message\\\":\\\" 6094 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0215 20:17:07.579544 6094 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:07.579626 6094 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0215 20:17:07.579723 6094 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:07.579765 6094 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0215 20:17:07.579782 6094 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0215 20:17:07.579812 6094 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:07.579822 6094 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:07.579833 6094 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:07.579789 6094 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:07.579835 6094 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0215 20:17:07.579863 6094 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:07.579886 6094 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0215 20:17:07.579934 6094 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0215 20:17:07.579974 6094 factory.go:656] Stopping watch factory\\\\nI0215 20:17:07.580011 6094 ovnkube.go:599] Stopped ovnkube\\\\nI02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.086784 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.110351 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.132913 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.157093 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.157152 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.157169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.157194 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.157213 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.158395 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.178470 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.204089 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.220363 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.236684 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:13Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.260203 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.260313 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.260334 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.260401 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.260420 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.364294 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:13 crc kubenswrapper[4735]: E0215 20:17:13.364501 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:13 crc kubenswrapper[4735]: E0215 20:17:13.364574 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs podName:96681837-f2a3-42d2-9653-a3997f396291 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:17.364547572 +0000 UTC m=+45.230563235 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs") pod "network-metrics-daemon-7vfs4" (UID: "96681837-f2a3-42d2-9653-a3997f396291") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.365291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.365637 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.365815 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.365991 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.366057 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.469379 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.469511 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.469536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.469570 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.469594 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.573193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.573350 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.573371 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.573445 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.573478 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.677994 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.678048 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.678064 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.678092 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.678106 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.781308 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.781383 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.781400 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.781430 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.781447 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.837856 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 03:44:26.851640928 +0000 UTC Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.884909 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.885023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.885042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.885070 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.885090 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.885793 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.885826 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.885856 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.885805 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:13 crc kubenswrapper[4735]: E0215 20:17:13.886022 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:13 crc kubenswrapper[4735]: E0215 20:17:13.886217 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:13 crc kubenswrapper[4735]: E0215 20:17:13.886471 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:13 crc kubenswrapper[4735]: E0215 20:17:13.886681 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.990566 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.990662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.990679 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.990703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:13 crc kubenswrapper[4735]: I0215 20:17:13.990720 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:13Z","lastTransitionTime":"2026-02-15T20:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.094799 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.094861 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.094883 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.094914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.094937 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:14Z","lastTransitionTime":"2026-02-15T20:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.198038 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.198163 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.198193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.198231 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.198266 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:14Z","lastTransitionTime":"2026-02-15T20:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.301366 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.301437 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.301456 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.301485 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.301504 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:14Z","lastTransitionTime":"2026-02-15T20:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.404367 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.404430 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.404442 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.404466 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.404477 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:14Z","lastTransitionTime":"2026-02-15T20:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.507328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.507401 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.507424 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.507455 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.507480 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:14Z","lastTransitionTime":"2026-02-15T20:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.611182 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.611313 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.611396 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.611430 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.611457 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:14Z","lastTransitionTime":"2026-02-15T20:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.717323 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.717378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.717403 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.717429 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.717446 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:14Z","lastTransitionTime":"2026-02-15T20:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.820092 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.820127 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.820135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.820153 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.820163 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:14Z","lastTransitionTime":"2026-02-15T20:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.839059 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 19:07:01.708890224 +0000 UTC Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.922906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.923009 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.923028 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.923055 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:14 crc kubenswrapper[4735]: I0215 20:17:14.923073 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:14Z","lastTransitionTime":"2026-02-15T20:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.026805 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.026881 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.026899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.026926 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.026968 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.130655 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.130772 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.130798 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.130836 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.130864 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.235378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.235449 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.235469 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.235496 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.235517 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.339673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.340211 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.340232 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.340261 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.340281 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.444414 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.444487 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.444510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.444543 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.444568 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.547737 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.547811 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.547837 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.547878 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.547902 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.650638 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.650708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.650727 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.650755 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.650775 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.754314 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.754380 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.754399 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.754452 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.754472 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.840057 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 11:53:42.975759491 +0000 UTC Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.857837 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.857895 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.857915 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.857995 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.858022 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.886630 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.886713 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.886788 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.886631 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:15 crc kubenswrapper[4735]: E0215 20:17:15.886898 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:15 crc kubenswrapper[4735]: E0215 20:17:15.887063 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:15 crc kubenswrapper[4735]: E0215 20:17:15.887226 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:15 crc kubenswrapper[4735]: E0215 20:17:15.887364 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.961580 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.961635 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.961651 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.961676 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:15 crc kubenswrapper[4735]: I0215 20:17:15.961694 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:15Z","lastTransitionTime":"2026-02-15T20:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.064744 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.064812 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.064829 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.064855 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.064875 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:16Z","lastTransitionTime":"2026-02-15T20:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.168148 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.168232 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.168322 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.168360 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.168382 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:16Z","lastTransitionTime":"2026-02-15T20:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.272035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.272094 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.272110 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.272135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.272190 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:16Z","lastTransitionTime":"2026-02-15T20:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.375841 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.375903 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.375924 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.375981 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.376002 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:16Z","lastTransitionTime":"2026-02-15T20:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.479458 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.479540 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.479565 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.479596 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.479618 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:16Z","lastTransitionTime":"2026-02-15T20:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.588204 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.588260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.588278 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.588305 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.588324 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:16Z","lastTransitionTime":"2026-02-15T20:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.692612 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.692684 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.692702 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.692733 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.692751 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:16Z","lastTransitionTime":"2026-02-15T20:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.796735 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.796785 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.796797 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.796816 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.796827 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:16Z","lastTransitionTime":"2026-02-15T20:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.840939 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 16:55:46.231670965 +0000 UTC Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.899558 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.899614 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.899632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.899659 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:16 crc kubenswrapper[4735]: I0215 20:17:16.899677 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:16Z","lastTransitionTime":"2026-02-15T20:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.003122 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.003161 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.003170 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.003188 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.003198 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.106839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.106892 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.106907 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.106932 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.106982 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.210301 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.210373 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.210399 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.210431 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.210454 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.314260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.314360 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.314379 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.314406 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.314427 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.420400 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:17 crc kubenswrapper[4735]: E0215 20:17:17.420747 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:17 crc kubenswrapper[4735]: E0215 20:17:17.420854 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs podName:96681837-f2a3-42d2-9653-a3997f396291 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:25.420822464 +0000 UTC m=+53.286838117 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs") pod "network-metrics-daemon-7vfs4" (UID: "96681837-f2a3-42d2-9653-a3997f396291") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.424628 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.424716 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.424738 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.424769 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.424796 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.528441 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.528502 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.528525 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.528552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.528573 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.632076 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.632176 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.632197 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.632224 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.632243 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.735394 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.735480 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.735501 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.735531 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.735555 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.838800 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.838854 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.838872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.838897 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.838913 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.841809 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 14:39:21.363879472 +0000 UTC Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.885890 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:17 crc kubenswrapper[4735]: E0215 20:17:17.886110 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.886437 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.886454 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.886525 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:17 crc kubenswrapper[4735]: E0215 20:17:17.886980 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:17 crc kubenswrapper[4735]: E0215 20:17:17.887076 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:17 crc kubenswrapper[4735]: E0215 20:17:17.887233 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.942721 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.942779 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.942796 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.942822 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:17 crc kubenswrapper[4735]: I0215 20:17:17.942841 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:17Z","lastTransitionTime":"2026-02-15T20:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.046393 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.046459 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.046482 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.046510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.046529 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.149343 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.149668 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.149833 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.150022 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.150182 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.254621 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.255071 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.255236 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.255391 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.255591 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.359032 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.359090 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.359113 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.359138 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.359156 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.462392 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.462460 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.462480 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.462510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.462531 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.566285 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.566699 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.567008 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.567200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.567485 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.672068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.672135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.672155 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.672183 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.672202 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.776499 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.776885 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.777187 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.777273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.777301 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.842861 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 20:10:38.91663491 +0000 UTC Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.881545 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.881611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.881631 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.881667 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.881694 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.985621 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.985690 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.985707 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.985734 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:18 crc kubenswrapper[4735]: I0215 20:17:18.985758 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:18Z","lastTransitionTime":"2026-02-15T20:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.089182 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.089232 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.089250 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.089275 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.089292 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:19Z","lastTransitionTime":"2026-02-15T20:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.192211 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.192276 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.192293 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.192318 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.192334 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:19Z","lastTransitionTime":"2026-02-15T20:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.295735 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.295794 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.295812 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.295839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.295856 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:19Z","lastTransitionTime":"2026-02-15T20:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.398761 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.398829 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.398848 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.398875 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.398895 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:19Z","lastTransitionTime":"2026-02-15T20:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.501897 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.501986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.502008 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.502034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.502053 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:19Z","lastTransitionTime":"2026-02-15T20:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.604801 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.604868 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.604885 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.604911 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.604980 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:19Z","lastTransitionTime":"2026-02-15T20:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.708250 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.708328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.708354 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.708390 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.708414 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:19Z","lastTransitionTime":"2026-02-15T20:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.812383 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.812448 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.812465 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.812493 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.812511 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:19Z","lastTransitionTime":"2026-02-15T20:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.843751 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 13:13:16.905656656 +0000 UTC Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.888185 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:19 crc kubenswrapper[4735]: E0215 20:17:19.888434 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.888614 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:19 crc kubenswrapper[4735]: E0215 20:17:19.888796 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.888826 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.888905 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:19 crc kubenswrapper[4735]: E0215 20:17:19.888980 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:19 crc kubenswrapper[4735]: E0215 20:17:19.889123 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.916354 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.916443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.916464 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.916497 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:19 crc kubenswrapper[4735]: I0215 20:17:19.916523 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:19Z","lastTransitionTime":"2026-02-15T20:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.020827 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.020897 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.020910 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.020931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.020974 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.030650 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.030723 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.030742 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.030770 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.030793 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: E0215 20:17:20.051989 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:20Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.059059 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.059140 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.059159 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.059186 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.059204 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: E0215 20:17:20.079325 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:20Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.085004 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.085076 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.085100 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.085130 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.085151 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: E0215 20:17:20.105312 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:20Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.110362 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.110416 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.110438 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.110469 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.110492 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: E0215 20:17:20.131660 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:20Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.137922 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.137993 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.138007 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.138031 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.138046 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: E0215 20:17:20.159598 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:20Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:20 crc kubenswrapper[4735]: E0215 20:17:20.159822 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.163070 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.163129 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.163148 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.163177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.163194 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.266018 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.266115 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.266136 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.266198 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.266219 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.368680 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.368753 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.368773 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.368802 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.368820 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.472071 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.472142 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.472161 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.472190 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.472210 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.574934 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.575032 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.575052 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.575078 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.575095 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.679192 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.679252 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.679268 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.679297 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.679314 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.782029 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.782118 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.782141 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.782166 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.782185 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.844096 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 03:28:17.411495057 +0000 UTC Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.885552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.885617 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.885638 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.885668 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.885695 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.989603 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.989671 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.989690 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.989716 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:20 crc kubenswrapper[4735]: I0215 20:17:20.989734 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:20Z","lastTransitionTime":"2026-02-15T20:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.093291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.093350 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.093371 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.093398 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.093415 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:21Z","lastTransitionTime":"2026-02-15T20:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.196728 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.196811 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.196839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.196875 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.196897 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:21Z","lastTransitionTime":"2026-02-15T20:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.300235 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.300291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.300309 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.300336 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.300354 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:21Z","lastTransitionTime":"2026-02-15T20:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.409653 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.409715 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.409734 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.409763 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.409782 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:21Z","lastTransitionTime":"2026-02-15T20:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.512763 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.512832 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.512852 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.512879 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.512898 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:21Z","lastTransitionTime":"2026-02-15T20:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.616552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.616615 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.616633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.616663 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.616685 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:21Z","lastTransitionTime":"2026-02-15T20:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.719899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.719943 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.719995 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.720014 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.720028 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:21Z","lastTransitionTime":"2026-02-15T20:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.823137 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.823179 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.823196 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.823218 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.823235 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:21Z","lastTransitionTime":"2026-02-15T20:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.844265 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 18:20:04.802911493 +0000 UTC Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.886699 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:21 crc kubenswrapper[4735]: E0215 20:17:21.886854 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.887176 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:21 crc kubenswrapper[4735]: E0215 20:17:21.887279 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.887370 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.887510 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:21 crc kubenswrapper[4735]: E0215 20:17:21.887561 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:21 crc kubenswrapper[4735]: E0215 20:17:21.887797 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.925657 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.925720 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.925739 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.925762 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:21 crc kubenswrapper[4735]: I0215 20:17:21.925782 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:21Z","lastTransitionTime":"2026-02-15T20:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.029307 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.029371 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.029390 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.029418 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.029443 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.131904 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.132007 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.132034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.132062 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.132081 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.235559 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.235617 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.235641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.235670 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.235693 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.339032 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.339077 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.339094 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.339116 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.339135 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.442547 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.442899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.443114 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.443273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.443422 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.547162 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.547256 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.547276 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.547304 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.547323 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.650542 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.650602 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.650622 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.650653 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.650677 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.754268 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.754351 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.754374 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.754406 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.754459 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.845464 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 03:33:27.610312751 +0000 UTC Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.856593 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.856642 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.856661 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.856686 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.856708 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.887526 4735 scope.go:117] "RemoveContainer" containerID="2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.916997 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:22Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.940745 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:22Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.963102 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.963182 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.963206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.963240 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.963267 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:22Z","lastTransitionTime":"2026-02-15T20:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.969314 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:22Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:22 crc kubenswrapper[4735]: I0215 20:17:22.990068 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:22Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.010260 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.026419 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.051301 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.066322 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.066370 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.066386 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.066408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.066423 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.072636 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.090508 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.108014 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.123991 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.151383 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.168609 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.168687 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.168716 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.168747 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.168771 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.170741 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.187583 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.217022 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb5cf7a37da3119a39b05ca6a1b2650376f3c880d60f36da9d74080195e76e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:05Z\\\",\\\"message\\\":\\\"5 20:17:05.412196 5957 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0215 20:17:05.412377 5957 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412698 5957 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.412872 5957 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:05.413376 5957 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:05.413422 5957 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:05.413459 5957 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:05.413469 5957 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:05.413473 5957 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:05.413492 5957 factory.go:656] Stopping watch factory\\\\nI0215 20:17:05.413514 5957 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:05.413528 5957 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:05.413539 5957 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:07Z\\\",\\\"message\\\":\\\" 6094 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0215 20:17:07.579544 6094 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:07.579626 6094 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0215 20:17:07.579723 6094 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:07.579765 6094 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0215 20:17:07.579782 6094 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0215 20:17:07.579812 6094 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:07.579822 6094 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:07.579833 6094 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:07.579789 6094 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:07.579835 6094 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0215 20:17:07.579863 6094 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:07.579886 6094 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0215 20:17:07.579934 6094 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0215 20:17:07.579974 6094 factory.go:656] Stopping watch factory\\\\nI0215 20:17:07.580011 6094 ovnkube.go:599] Stopped ovnkube\\\\nI02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.230841 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.249335 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.265638 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.271133 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.271188 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.271200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.271220 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.271235 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.284528 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.301004 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.317328 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.334688 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.352063 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.356373 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/1.log" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.360575 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436"} Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.360769 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.369023 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.373679 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.373717 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.373729 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.373747 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.373758 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.391306 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.408373 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.425363 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.444966 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.468475 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.475879 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.475902 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.475912 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.475969 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.475982 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.492038 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.520344 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:07Z\\\",\\\"message\\\":\\\" 6094 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0215 20:17:07.579544 6094 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:07.579626 6094 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0215 20:17:07.579723 6094 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:07.579765 6094 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0215 20:17:07.579782 6094 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0215 20:17:07.579812 6094 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:07.579822 6094 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:07.579833 6094 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:07.579789 6094 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:07.579835 6094 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0215 20:17:07.579863 6094 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:07.579886 6094 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0215 20:17:07.579934 6094 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0215 20:17:07.579974 6094 factory.go:656] Stopping watch factory\\\\nI0215 20:17:07.580011 6094 ovnkube.go:599] Stopped ovnkube\\\\nI02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.538737 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.561227 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.578725 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.578767 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.578779 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.578802 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.578817 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.591745 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.612286 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.625064 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.643705 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.662305 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.681076 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.681126 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.681137 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.681157 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.681167 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.684014 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:07Z\\\",\\\"message\\\":\\\" 6094 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0215 20:17:07.579544 6094 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:07.579626 6094 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0215 20:17:07.579723 6094 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:07.579765 6094 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0215 20:17:07.579782 6094 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0215 20:17:07.579812 6094 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:07.579822 6094 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:07.579833 6094 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:07.579789 6094 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:07.579835 6094 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0215 20:17:07.579863 6094 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:07.579886 6094 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0215 20:17:07.579934 6094 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0215 20:17:07.579974 6094 factory.go:656] Stopping watch factory\\\\nI0215 20:17:07.580011 6094 ovnkube.go:599] Stopped ovnkube\\\\nI02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.696053 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.711529 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.725939 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.742160 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.758675 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.772517 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.783967 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.783997 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.784010 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.784027 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.784037 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.787641 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.800242 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.812902 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:23Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.846454 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 09:19:23.190092075 +0000 UTC Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.885685 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.885735 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.885804 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.885880 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:23 crc kubenswrapper[4735]: E0215 20:17:23.885882 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:23 crc kubenswrapper[4735]: E0215 20:17:23.886032 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:23 crc kubenswrapper[4735]: E0215 20:17:23.886131 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.886538 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.886592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.886603 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.886626 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.886640 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:23 crc kubenswrapper[4735]: E0215 20:17:23.886642 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.989776 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.990084 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.990179 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.990312 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:23 crc kubenswrapper[4735]: I0215 20:17:23.990400 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:23Z","lastTransitionTime":"2026-02-15T20:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.093435 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.093602 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.093690 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.093774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.093856 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:24Z","lastTransitionTime":"2026-02-15T20:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.198059 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.198131 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.198155 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.198200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.198220 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:24Z","lastTransitionTime":"2026-02-15T20:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.302158 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.302244 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.302265 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.302295 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.302315 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:24Z","lastTransitionTime":"2026-02-15T20:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.369528 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/2.log" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.370986 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/1.log" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.376284 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436" exitCode=1 Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.376377 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.376687 4735 scope.go:117] "RemoveContainer" containerID="2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.381079 4735 scope.go:117] "RemoveContainer" containerID="0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436" Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.381593 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.406012 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.406068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.406088 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.406116 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.406134 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:24Z","lastTransitionTime":"2026-02-15T20:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.408257 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.435072 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.457403 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.480150 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.498764 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.509314 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.509382 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.509402 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.509431 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.509451 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:24Z","lastTransitionTime":"2026-02-15T20:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.521810 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.547494 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.571504 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.591762 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.610868 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.612863 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.613113 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.613260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.613396 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.613516 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:24Z","lastTransitionTime":"2026-02-15T20:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.631228 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.651919 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.687361 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2891757dd197fa5e6043333aece24e5751673b08fd7a176b8568be3e66232dc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:07Z\\\",\\\"message\\\":\\\" 6094 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0215 20:17:07.579544 6094 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:07.579626 6094 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0215 20:17:07.579723 6094 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0215 20:17:07.579765 6094 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0215 20:17:07.579782 6094 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0215 20:17:07.579812 6094 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0215 20:17:07.579822 6094 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:07.579833 6094 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:07.579789 6094 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:07.579835 6094 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0215 20:17:07.579863 6094 handler.go:208] Removed *v1.Node event handler 2\\\\nI0215 20:17:07.579886 6094 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0215 20:17:07.579934 6094 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0215 20:17:07.579974 6094 factory.go:656] Stopping watch factory\\\\nI0215 20:17:07.580011 6094 ovnkube.go:599] Stopped ovnkube\\\\nI02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.709871 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.717144 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.717213 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.717233 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.717260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.717277 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:24Z","lastTransitionTime":"2026-02-15T20:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.730772 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.742860 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.743112 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.743207 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.743410 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.743569 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:56.743539527 +0000 UTC m=+84.609555190 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.743642 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.743658 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:17:56.743641259 +0000 UTC m=+84.609656922 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.743719 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:56.743694311 +0000 UTC m=+84.609709934 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.747868 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:24Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.821155 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.821347 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.821371 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.821398 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.821417 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:24Z","lastTransitionTime":"2026-02-15T20:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.844113 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.844260 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.844387 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.844432 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.844455 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.844526 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.844547 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:56.84451512 +0000 UTC m=+84.710530783 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.844561 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.844585 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:24 crc kubenswrapper[4735]: E0215 20:17:24.844676 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:56.844644063 +0000 UTC m=+84.710659906 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.846989 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 04:07:51.360683727 +0000 UTC Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.926180 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.926230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.926246 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.926273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.926290 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:24Z","lastTransitionTime":"2026-02-15T20:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:24 crc kubenswrapper[4735]: I0215 20:17:24.928440 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.030154 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.030238 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.030257 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.030290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.030315 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.133754 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.133828 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.133849 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.133876 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.133895 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.237497 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.237564 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.237583 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.237610 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.237628 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.341740 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.341793 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.341810 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.341834 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.341850 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.382478 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/2.log" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.388017 4735 scope.go:117] "RemoveContainer" containerID="0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436" Feb 15 20:17:25 crc kubenswrapper[4735]: E0215 20:17:25.388299 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.410670 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.431923 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.444696 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.444752 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.444770 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.444802 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.444823 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.449535 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.451477 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:25 crc kubenswrapper[4735]: E0215 20:17:25.451652 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:25 crc kubenswrapper[4735]: E0215 20:17:25.451733 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs podName:96681837-f2a3-42d2-9653-a3997f396291 nodeName:}" failed. No retries permitted until 2026-02-15 20:17:41.451708184 +0000 UTC m=+69.317723847 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs") pod "network-metrics-daemon-7vfs4" (UID: "96681837-f2a3-42d2-9653-a3997f396291") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.467755 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.485751 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.500917 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.517818 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.536867 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.548809 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.548869 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.548887 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.548915 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.548933 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.567285 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.586573 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.608270 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.629649 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.651985 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.652082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.652151 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.652189 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.652246 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.654279 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.669907 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.687350 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.703551 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:25Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.756597 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.756686 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.756715 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.756749 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.756772 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.847814 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 16:41:00.427794127 +0000 UTC Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.861568 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.861642 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.861671 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.861704 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.861724 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.886163 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.886268 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.886343 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.886236 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:25 crc kubenswrapper[4735]: E0215 20:17:25.886464 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:25 crc kubenswrapper[4735]: E0215 20:17:25.886651 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:25 crc kubenswrapper[4735]: E0215 20:17:25.886859 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:25 crc kubenswrapper[4735]: E0215 20:17:25.887078 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.965186 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.965268 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.965290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.965322 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:25 crc kubenswrapper[4735]: I0215 20:17:25.965341 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:25Z","lastTransitionTime":"2026-02-15T20:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.068463 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.068534 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.068552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.068578 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.068601 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:26Z","lastTransitionTime":"2026-02-15T20:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.172716 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.172913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.172933 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.172994 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.173014 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:26Z","lastTransitionTime":"2026-02-15T20:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.277093 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.277166 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.277197 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.277226 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.277247 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:26Z","lastTransitionTime":"2026-02-15T20:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.381798 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.381883 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.381907 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.381937 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.381989 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:26Z","lastTransitionTime":"2026-02-15T20:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.488070 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.488825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.488902 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.489037 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.489066 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:26Z","lastTransitionTime":"2026-02-15T20:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.592572 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.592648 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.592671 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.592703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.592723 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:26Z","lastTransitionTime":"2026-02-15T20:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.696730 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.696801 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.696824 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.696851 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.696872 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:26Z","lastTransitionTime":"2026-02-15T20:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.800616 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.800705 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.800723 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.800751 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.800771 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:26Z","lastTransitionTime":"2026-02-15T20:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.849042 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 03:46:46.309228024 +0000 UTC Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.904329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.904431 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.904454 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.904486 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:26 crc kubenswrapper[4735]: I0215 20:17:26.904511 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:26Z","lastTransitionTime":"2026-02-15T20:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.008854 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.008931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.008991 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.009026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.009052 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.112386 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.112430 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.112443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.112462 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.112476 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.215812 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.215864 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.215875 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.215895 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.215911 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.318760 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.318829 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.318849 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.318878 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.318899 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.323379 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.334417 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.344835 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.363798 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.376059 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.386850 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.409253 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.422464 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.422548 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.422568 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.422626 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.422643 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.429542 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.451667 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.487828 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.503662 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.520142 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.531728 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.531795 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.531808 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.531831 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.531864 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.539226 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.556184 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.570147 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.585397 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.596687 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.613018 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:27Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.635124 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.635178 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.635197 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.635222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.635241 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.739373 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.739438 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.739459 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.739486 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.739508 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.842565 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.842654 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.842676 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.842732 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.842754 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.849567 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 05:50:59.383735941 +0000 UTC Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.886304 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.886342 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.886371 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.886472 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:27 crc kubenswrapper[4735]: E0215 20:17:27.886688 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:27 crc kubenswrapper[4735]: E0215 20:17:27.887336 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:27 crc kubenswrapper[4735]: E0215 20:17:27.887526 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:27 crc kubenswrapper[4735]: E0215 20:17:27.887690 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.950298 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.950500 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.950574 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.950617 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:27 crc kubenswrapper[4735]: I0215 20:17:27.950702 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:27Z","lastTransitionTime":"2026-02-15T20:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.059371 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.059448 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.059466 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.059499 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.059519 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.163401 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.163455 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.163473 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.163499 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.163517 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.266740 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.266800 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.266817 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.266845 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.266863 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.370639 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.370692 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.370709 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.370733 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.370751 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.474475 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.474532 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.474552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.474577 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.474594 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.578558 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.578656 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.578676 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.578699 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.578753 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.682125 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.682192 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.682216 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.682247 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.682270 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.786098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.786157 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.786183 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.786211 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.786232 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.850243 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 10:54:00.147988815 +0000 UTC Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.889138 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.889204 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.889226 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.889257 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.889278 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.993079 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.993154 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.993177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.993223 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:28 crc kubenswrapper[4735]: I0215 20:17:28.993250 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:28Z","lastTransitionTime":"2026-02-15T20:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.096519 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.096575 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.096592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.096617 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.096634 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:29Z","lastTransitionTime":"2026-02-15T20:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.199802 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.199888 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.199914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.199997 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.200040 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:29Z","lastTransitionTime":"2026-02-15T20:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.302523 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.302585 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.302606 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.302635 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.302662 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:29Z","lastTransitionTime":"2026-02-15T20:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.405052 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.405173 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.405193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.405222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.405240 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:29Z","lastTransitionTime":"2026-02-15T20:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.508617 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.508685 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.508708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.508736 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.508758 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:29Z","lastTransitionTime":"2026-02-15T20:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.611615 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.611680 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.611700 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.611726 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.611743 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:29Z","lastTransitionTime":"2026-02-15T20:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.714658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.714727 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.714750 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.714782 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.714805 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:29Z","lastTransitionTime":"2026-02-15T20:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.818263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.818323 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.818340 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.818368 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.818386 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:29Z","lastTransitionTime":"2026-02-15T20:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.851273 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 03:06:41.92731653 +0000 UTC Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.886698 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.886787 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.886778 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:29 crc kubenswrapper[4735]: E0215 20:17:29.886859 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:29 crc kubenswrapper[4735]: E0215 20:17:29.887160 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:29 crc kubenswrapper[4735]: E0215 20:17:29.887290 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.887579 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:29 crc kubenswrapper[4735]: E0215 20:17:29.887739 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.920938 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.921013 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.921031 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.921053 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:29 crc kubenswrapper[4735]: I0215 20:17:29.921068 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:29Z","lastTransitionTime":"2026-02-15T20:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.024291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.024460 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.024478 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.024507 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.024524 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.127408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.127831 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.128026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.128222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.128372 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.231358 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.231419 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.231438 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.231463 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.231482 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.334334 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.334403 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.334428 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.334456 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.334476 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.437774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.438151 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.438177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.438209 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.438230 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.510755 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.510821 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.510844 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.510872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.510896 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: E0215 20:17:30.533381 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:30Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.540519 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.540592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.540613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.540643 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.540664 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: E0215 20:17:30.561093 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:30Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.566383 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.566455 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.566482 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.566513 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.566535 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: E0215 20:17:30.588392 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:30Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.593833 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.593887 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.593904 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.593929 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.593985 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: E0215 20:17:30.616099 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:30Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.622579 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.622659 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.622681 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.622716 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.622740 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: E0215 20:17:30.644243 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:30Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:30 crc kubenswrapper[4735]: E0215 20:17:30.644612 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.648408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.648487 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.648510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.648541 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.648563 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.752408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.752472 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.752490 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.752515 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.752536 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.851773 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 16:08:58.49319104 +0000 UTC Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.855578 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.855632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.855655 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.855686 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.855711 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.959325 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.959397 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.959415 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.959449 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:30 crc kubenswrapper[4735]: I0215 20:17:30.959467 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:30Z","lastTransitionTime":"2026-02-15T20:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.062167 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.062233 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.062255 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.062285 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.062308 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.166043 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.166111 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.166134 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.166164 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.166186 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.269363 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.269402 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.269419 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.269443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.269461 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.373074 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.373137 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.373154 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.373178 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.373195 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.476665 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.476730 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.476748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.476775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.476795 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.580130 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.580190 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.580206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.580230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.580247 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.684410 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.684467 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.684485 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.684511 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.684530 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.787221 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.787315 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.787342 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.787377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.787394 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.852363 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 13:22:20.831829324 +0000 UTC Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.886522 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.886682 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.886542 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.886693 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:31 crc kubenswrapper[4735]: E0215 20:17:31.886774 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:31 crc kubenswrapper[4735]: E0215 20:17:31.887021 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:31 crc kubenswrapper[4735]: E0215 20:17:31.887123 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:31 crc kubenswrapper[4735]: E0215 20:17:31.887283 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.891298 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.891395 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.891422 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.891452 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.891488 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.994845 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.994903 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.994998 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.995024 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:31 crc kubenswrapper[4735]: I0215 20:17:31.995077 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:31Z","lastTransitionTime":"2026-02-15T20:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.098762 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.098821 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.098839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.098863 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.098881 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:32Z","lastTransitionTime":"2026-02-15T20:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.201680 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.201744 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.201762 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.201789 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.201810 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:32Z","lastTransitionTime":"2026-02-15T20:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.305380 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.305446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.305464 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.305492 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.305510 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:32Z","lastTransitionTime":"2026-02-15T20:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.408866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.408931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.408974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.408999 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.409016 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:32Z","lastTransitionTime":"2026-02-15T20:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.511899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.512021 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.512048 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.512081 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.512108 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:32Z","lastTransitionTime":"2026-02-15T20:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.616710 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.616791 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.616809 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.616839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.616857 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:32Z","lastTransitionTime":"2026-02-15T20:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.720641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.720704 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.720721 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.720748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.720770 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:32Z","lastTransitionTime":"2026-02-15T20:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.825042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.825111 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.825129 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.825157 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.825178 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:32Z","lastTransitionTime":"2026-02-15T20:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.853399 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 10:40:04.370888986 +0000 UTC Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.910814 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:32Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.929303 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.929408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.929430 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.930184 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.930318 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:32Z","lastTransitionTime":"2026-02-15T20:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.932798 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:32Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.955875 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:32Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:32 crc kubenswrapper[4735]: I0215 20:17:32.990321 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:32Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.031303 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.033681 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.033765 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.033781 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.033821 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.033837 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.064804 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.083849 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.096590 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba92e95-6735-4b92-b51c-4358fe0bbea1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fc54fa7126011e839c783514b4d553d25a86a4175c053e84301561b281d19fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa568691661d11a2d6c03d89986f384f64101f0a2047d4865cc4a4b04fc2483c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740e7c6d10a65b9d2bdc9cba9d431aabb9c20e9bcce1fb2f40366a8a53a4f931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.110501 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.123025 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.134354 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.136537 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.136572 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.136603 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.136628 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.136641 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.150197 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.162832 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.179917 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.200690 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.213483 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.226817 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:33Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.240075 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.240136 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.240153 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.240227 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.240256 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.343108 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.343196 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.343220 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.343258 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.343288 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.447182 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.447269 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.447297 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.447333 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.447358 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.551209 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.551717 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.551740 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.551770 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.551789 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.654736 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.654826 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.654846 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.654870 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.654918 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.758389 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.758445 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.758465 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.758491 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.758510 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.853626 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 21:16:21.506100129 +0000 UTC Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.873774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.873852 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.873876 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.873910 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.873934 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.886126 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.886217 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.886314 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.886149 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:33 crc kubenswrapper[4735]: E0215 20:17:33.886326 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:33 crc kubenswrapper[4735]: E0215 20:17:33.886468 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:33 crc kubenswrapper[4735]: E0215 20:17:33.886581 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:33 crc kubenswrapper[4735]: E0215 20:17:33.886663 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.977159 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.977211 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.977230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.977254 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:33 crc kubenswrapper[4735]: I0215 20:17:33.977271 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:33Z","lastTransitionTime":"2026-02-15T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.081865 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.081927 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.081967 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.081991 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.082009 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:34Z","lastTransitionTime":"2026-02-15T20:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.185356 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.185424 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.185443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.185641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.185659 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:34Z","lastTransitionTime":"2026-02-15T20:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.289298 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.289377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.289397 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.289423 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.289443 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:34Z","lastTransitionTime":"2026-02-15T20:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.392531 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.392614 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.392631 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.392655 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.392673 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:34Z","lastTransitionTime":"2026-02-15T20:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.496128 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.496215 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.496258 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.496291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.496316 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:34Z","lastTransitionTime":"2026-02-15T20:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.600408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.600491 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.600533 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.600568 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.600589 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:34Z","lastTransitionTime":"2026-02-15T20:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.703888 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.704021 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.704042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.704068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.704086 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:34Z","lastTransitionTime":"2026-02-15T20:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.807506 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.807613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.807633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.807662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.807679 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:34Z","lastTransitionTime":"2026-02-15T20:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.854809 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 17:12:13.954073745 +0000 UTC Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.910879 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.910932 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.910975 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.911000 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:34 crc kubenswrapper[4735]: I0215 20:17:34.911018 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:34Z","lastTransitionTime":"2026-02-15T20:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.052453 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.052496 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.052508 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.052528 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.052542 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.155244 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.155292 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.155311 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.155338 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.155356 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.258346 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.258441 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.258463 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.258499 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.258523 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.361689 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.361750 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.361770 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.361798 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.361817 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.464935 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.464997 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.465008 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.465025 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.465037 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.568625 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.568692 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.568714 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.568743 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.568762 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.672376 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.672471 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.672496 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.672521 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.672540 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.776620 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.776680 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.776705 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.776736 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.776760 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.855512 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 16:16:10.284608845 +0000 UTC Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.881601 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.881680 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.881707 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.881742 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.881770 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.886496 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:35 crc kubenswrapper[4735]: E0215 20:17:35.886675 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.886982 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:35 crc kubenswrapper[4735]: E0215 20:17:35.887086 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.887283 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:35 crc kubenswrapper[4735]: E0215 20:17:35.887382 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.887581 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:35 crc kubenswrapper[4735]: E0215 20:17:35.887696 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.985424 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.985524 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.985547 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.985573 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:35 crc kubenswrapper[4735]: I0215 20:17:35.985592 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:35Z","lastTransitionTime":"2026-02-15T20:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.091298 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.091385 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.091411 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.091442 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.091486 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:36Z","lastTransitionTime":"2026-02-15T20:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.194740 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.194800 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.194816 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.194841 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.194860 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:36Z","lastTransitionTime":"2026-02-15T20:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.298573 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.298643 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.298661 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.298693 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.298712 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:36Z","lastTransitionTime":"2026-02-15T20:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.401120 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.401190 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.401207 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.401233 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.401251 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:36Z","lastTransitionTime":"2026-02-15T20:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.504073 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.504128 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.504143 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.504164 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.504178 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:36Z","lastTransitionTime":"2026-02-15T20:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.610408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.610457 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.610468 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.610487 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.610500 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:36Z","lastTransitionTime":"2026-02-15T20:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.713212 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.713248 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.713258 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.713273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.713281 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:36Z","lastTransitionTime":"2026-02-15T20:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.816374 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.816492 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.816504 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.816524 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.816536 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:36Z","lastTransitionTime":"2026-02-15T20:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.855615 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 23:11:22.974524491 +0000 UTC Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.918866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.918899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.918907 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.918922 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:36 crc kubenswrapper[4735]: I0215 20:17:36.918932 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:36Z","lastTransitionTime":"2026-02-15T20:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.024041 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.024082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.024093 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.024111 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.024122 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.126340 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.126384 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.126395 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.126413 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.126424 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.229173 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.229236 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.229256 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.229281 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.229297 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.331516 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.331558 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.331573 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.331591 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.331602 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.434242 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.434291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.434302 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.434321 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.434332 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.536555 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.536595 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.536605 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.536623 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.536632 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.640178 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.640283 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.640323 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.640366 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.640392 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.743519 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.743571 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.743589 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.743613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.743630 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.846842 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.846909 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.846927 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.846980 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.847002 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.856029 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 23:43:01.410846377 +0000 UTC Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.886813 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.886797 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.886819 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.886829 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:37 crc kubenswrapper[4735]: E0215 20:17:37.887102 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:37 crc kubenswrapper[4735]: E0215 20:17:37.887199 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:37 crc kubenswrapper[4735]: E0215 20:17:37.887310 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:37 crc kubenswrapper[4735]: E0215 20:17:37.887424 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.950674 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.950735 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.950753 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.950781 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:37 crc kubenswrapper[4735]: I0215 20:17:37.950802 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:37Z","lastTransitionTime":"2026-02-15T20:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.054001 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.054041 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.054080 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.054099 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.054112 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.156859 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.156889 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.156918 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.156935 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.156964 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.260139 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.260181 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.260192 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.260206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.260215 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.363028 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.363089 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.363106 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.363132 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.363158 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.465698 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.465738 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.465748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.465771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.465785 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.572808 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.572884 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.572905 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.572940 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.573001 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.677413 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.677476 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.677499 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.677531 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.677555 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.779862 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.779916 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.779932 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.779971 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.779988 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.857109 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 00:19:27.838291511 +0000 UTC Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.882393 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.882441 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.882450 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.882467 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.882477 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.886908 4735 scope.go:117] "RemoveContainer" containerID="0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436" Feb 15 20:17:38 crc kubenswrapper[4735]: E0215 20:17:38.887097 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.985328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.985377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.985388 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.985404 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:38 crc kubenswrapper[4735]: I0215 20:17:38.985414 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:38Z","lastTransitionTime":"2026-02-15T20:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.088277 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.088318 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.088328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.088357 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.088367 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:39Z","lastTransitionTime":"2026-02-15T20:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.191443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.191501 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.191519 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.191539 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.191550 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:39Z","lastTransitionTime":"2026-02-15T20:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.295447 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.295538 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.295564 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.295604 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.295630 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:39Z","lastTransitionTime":"2026-02-15T20:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.398322 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.398392 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.398413 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.398447 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.398468 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:39Z","lastTransitionTime":"2026-02-15T20:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.501703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.501752 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.501763 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.501780 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.501797 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:39Z","lastTransitionTime":"2026-02-15T20:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.604550 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.604618 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.604631 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.604650 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.604664 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:39Z","lastTransitionTime":"2026-02-15T20:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.706737 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.706781 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.706792 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.706808 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.706817 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:39Z","lastTransitionTime":"2026-02-15T20:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.810575 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.810632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.810645 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.810662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.810674 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:39Z","lastTransitionTime":"2026-02-15T20:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.858178 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 06:29:02.369036025 +0000 UTC Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.885700 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.885814 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:39 crc kubenswrapper[4735]: E0215 20:17:39.885944 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.886171 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.886207 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:39 crc kubenswrapper[4735]: E0215 20:17:39.886256 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:39 crc kubenswrapper[4735]: E0215 20:17:39.886567 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:39 crc kubenswrapper[4735]: E0215 20:17:39.886641 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.914536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.914583 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.914596 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.914613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:39 crc kubenswrapper[4735]: I0215 20:17:39.914624 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:39Z","lastTransitionTime":"2026-02-15T20:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.017481 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.017517 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.017529 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.017546 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.017560 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.120553 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.120583 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.120592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.120605 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.120615 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.223063 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.223117 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.223129 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.223148 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.223161 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.326190 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.326255 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.326270 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.326294 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.326310 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.429633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.429675 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.429684 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.429702 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.429714 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.532081 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.532123 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.532135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.532156 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.532169 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.636133 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.636208 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.636229 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.636260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.636282 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.711520 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.711592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.711605 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.711634 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.711648 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: E0215 20:17:40.734597 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:40Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.740775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.740824 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.740842 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.740870 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.740892 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: E0215 20:17:40.758935 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:40Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.764266 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.764329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.764349 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.764376 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.764394 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: E0215 20:17:40.779821 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:40Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.784157 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.784202 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.784218 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.784241 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.784256 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: E0215 20:17:40.797618 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:40Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.801519 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.801552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.801562 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.801591 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.801605 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: E0215 20:17:40.814637 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:40Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:40 crc kubenswrapper[4735]: E0215 20:17:40.814769 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.816787 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.816819 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.816830 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.816849 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.816860 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.859309 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 13:29:43.663592901 +0000 UTC Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.919849 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.919883 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.919892 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.919908 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:40 crc kubenswrapper[4735]: I0215 20:17:40.919920 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:40Z","lastTransitionTime":"2026-02-15T20:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.022290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.022361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.022381 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.022408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.022428 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.124533 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.124562 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.124570 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.124590 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.124599 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.227414 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.227465 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.227478 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.227498 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.227513 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.330963 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.331018 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.331027 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.331047 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.331060 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.433814 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.433865 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.433874 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.433889 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.433898 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.536311 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.536366 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.536380 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.536404 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.536422 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.543830 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:41 crc kubenswrapper[4735]: E0215 20:17:41.544014 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:41 crc kubenswrapper[4735]: E0215 20:17:41.544097 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs podName:96681837-f2a3-42d2-9653-a3997f396291 nodeName:}" failed. No retries permitted until 2026-02-15 20:18:13.54407102 +0000 UTC m=+101.410086653 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs") pod "network-metrics-daemon-7vfs4" (UID: "96681837-f2a3-42d2-9653-a3997f396291") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.639091 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.639140 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.639151 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.639164 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.639173 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.742469 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.742540 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.742561 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.742594 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.742613 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.845562 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.845641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.845661 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.845696 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.845715 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.860305 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 05:07:24.059658617 +0000 UTC Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.886790 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.886870 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.886812 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.886803 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:41 crc kubenswrapper[4735]: E0215 20:17:41.886977 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:41 crc kubenswrapper[4735]: E0215 20:17:41.887118 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:41 crc kubenswrapper[4735]: E0215 20:17:41.887209 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:41 crc kubenswrapper[4735]: E0215 20:17:41.887335 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.948872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.948932 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.948956 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.948977 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:41 crc kubenswrapper[4735]: I0215 20:17:41.948989 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:41Z","lastTransitionTime":"2026-02-15T20:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.052178 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.052225 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.052249 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.052270 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.052281 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.156128 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.156200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.156220 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.156251 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.156273 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.259416 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.259501 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.259530 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.259564 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.259588 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.362823 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.362876 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.362887 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.362905 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.362919 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.453258 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/0.log" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.453341 4735 generic.go:334] "Generic (PLEG): container finished" podID="a20277cf-39a2-453f-ac6d-fb28346b3358" containerID="4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9" exitCode=1 Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.453384 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wfjmx" event={"ID":"a20277cf-39a2-453f-ac6d-fb28346b3358","Type":"ContainerDied","Data":"4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.453933 4735 scope.go:117] "RemoveContainer" containerID="4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.467147 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.467277 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.467305 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.467330 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.467350 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.473186 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.492922 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.503763 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.518012 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.534051 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.554719 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.568101 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.570771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.570834 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.570846 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.570866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.570878 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.584148 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.601136 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.622406 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.640454 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.657330 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.673562 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.673794 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.673863 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.673924 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.674010 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.675979 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"2026-02-15T20:16:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd\\\\n2026-02-15T20:16:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd to /host/opt/cni/bin/\\\\n2026-02-15T20:16:57Z [verbose] multus-daemon started\\\\n2026-02-15T20:16:57Z [verbose] Readiness Indicator file check\\\\n2026-02-15T20:17:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.691559 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.704881 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.718850 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba92e95-6735-4b92-b51c-4358fe0bbea1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fc54fa7126011e839c783514b4d553d25a86a4175c053e84301561b281d19fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa568691661d11a2d6c03d89986f384f64101f0a2047d4865cc4a4b04fc2483c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740e7c6d10a65b9d2bdc9cba9d431aabb9c20e9bcce1fb2f40366a8a53a4f931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.732434 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.776986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.777113 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.777170 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.777227 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.777286 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.860998 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 14:03:50.331180473 +0000 UTC Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.881212 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.881279 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.881300 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.881326 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.881343 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.908038 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.927878 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.941002 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.955363 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.969433 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.981469 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.983731 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.983776 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.983789 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.983810 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:42 crc kubenswrapper[4735]: I0215 20:17:42.983853 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:42Z","lastTransitionTime":"2026-02-15T20:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.000373 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:42Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.021783 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.034342 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.048657 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.061419 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.084937 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.085872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.085920 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.085933 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.085969 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.085983 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:43Z","lastTransitionTime":"2026-02-15T20:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.099012 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba92e95-6735-4b92-b51c-4358fe0bbea1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fc54fa7126011e839c783514b4d553d25a86a4175c053e84301561b281d19fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa568691661d11a2d6c03d89986f384f64101f0a2047d4865cc4a4b04fc2483c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740e7c6d10a65b9d2bdc9cba9d431aabb9c20e9bcce1fb2f40366a8a53a4f931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.112854 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.126681 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"2026-02-15T20:16:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd\\\\n2026-02-15T20:16:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd to /host/opt/cni/bin/\\\\n2026-02-15T20:16:57Z [verbose] multus-daemon started\\\\n2026-02-15T20:16:57Z [verbose] Readiness Indicator file check\\\\n2026-02-15T20:17:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.138089 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.153165 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.188737 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.188807 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.188822 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.188846 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.188861 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:43Z","lastTransitionTime":"2026-02-15T20:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.292076 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.292154 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.292170 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.292192 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.292208 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:43Z","lastTransitionTime":"2026-02-15T20:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.395459 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.395794 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.395872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.395976 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.396048 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:43Z","lastTransitionTime":"2026-02-15T20:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.460551 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/0.log" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.460684 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wfjmx" event={"ID":"a20277cf-39a2-453f-ac6d-fb28346b3358","Type":"ContainerStarted","Data":"17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.479371 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.492503 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.498769 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.498820 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.498838 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.498864 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.498882 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:43Z","lastTransitionTime":"2026-02-15T20:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.506698 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.520095 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.534495 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.547714 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.567191 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.588067 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.600180 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.601746 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.601905 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.602209 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.602326 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.602406 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:43Z","lastTransitionTime":"2026-02-15T20:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.617361 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.632603 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.649782 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.664367 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba92e95-6735-4b92-b51c-4358fe0bbea1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fc54fa7126011e839c783514b4d553d25a86a4175c053e84301561b281d19fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa568691661d11a2d6c03d89986f384f64101f0a2047d4865cc4a4b04fc2483c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740e7c6d10a65b9d2bdc9cba9d431aabb9c20e9bcce1fb2f40366a8a53a4f931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.677713 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.692286 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"2026-02-15T20:16:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd\\\\n2026-02-15T20:16:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd to /host/opt/cni/bin/\\\\n2026-02-15T20:16:57Z [verbose] multus-daemon started\\\\n2026-02-15T20:16:57Z [verbose] Readiness Indicator file check\\\\n2026-02-15T20:17:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.703190 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.705251 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.705289 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.705302 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.705322 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.705336 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:43Z","lastTransitionTime":"2026-02-15T20:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.717334 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:43Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.808843 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.809592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.809691 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.809797 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.809881 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:43Z","lastTransitionTime":"2026-02-15T20:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.861684 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 17:58:53.14283578 +0000 UTC Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.886224 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:43 crc kubenswrapper[4735]: E0215 20:17:43.886392 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.886638 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:43 crc kubenswrapper[4735]: E0215 20:17:43.886709 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.886646 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:43 crc kubenswrapper[4735]: E0215 20:17:43.887195 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.887339 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:43 crc kubenswrapper[4735]: E0215 20:17:43.887605 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.913006 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.913049 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.913061 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.913082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:43 crc kubenswrapper[4735]: I0215 20:17:43.913098 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:43Z","lastTransitionTime":"2026-02-15T20:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.015914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.015992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.016005 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.016024 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.016035 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.118779 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.119082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.119145 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.119205 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.119284 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.226632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.226916 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.227016 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.227085 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.227149 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.330968 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.331026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.331041 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.331066 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.331082 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.434840 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.434896 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.434906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.434926 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.434939 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.537451 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.537516 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.537535 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.537561 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.537582 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.640246 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.640285 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.640298 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.640316 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.640333 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.743013 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.743041 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.743049 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.743062 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.743071 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.845929 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.845992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.846069 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.846089 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.846098 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.862360 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 15:20:32.245994175 +0000 UTC Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.947708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.947934 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.948062 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.948126 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:44 crc kubenswrapper[4735]: I0215 20:17:44.948179 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:44Z","lastTransitionTime":"2026-02-15T20:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.051026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.051561 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.051624 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.051684 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.051738 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.154249 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.154295 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.154305 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.154322 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.154333 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.257251 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.257538 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.257643 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.257709 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.257772 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.363182 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.363443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.363521 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.363632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.363706 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.466527 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.466669 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.466869 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.467067 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.467225 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.571428 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.571663 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.571724 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.571783 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.571835 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.675837 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.676077 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.676149 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.676213 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.676290 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.779506 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.779906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.780176 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.780608 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.781003 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.863659 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 18:21:03.122089523 +0000 UTC Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.884458 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.884527 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.884858 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.884889 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.884910 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.886326 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.886403 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.886499 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.886351 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:45 crc kubenswrapper[4735]: E0215 20:17:45.886534 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:45 crc kubenswrapper[4735]: E0215 20:17:45.886679 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:45 crc kubenswrapper[4735]: E0215 20:17:45.886820 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:45 crc kubenswrapper[4735]: E0215 20:17:45.887214 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.988238 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.988292 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.988310 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.988335 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:45 crc kubenswrapper[4735]: I0215 20:17:45.988354 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:45Z","lastTransitionTime":"2026-02-15T20:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.092350 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.092404 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.092416 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.092436 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.092451 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:46Z","lastTransitionTime":"2026-02-15T20:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.194779 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.194818 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.194835 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.194858 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.194873 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:46Z","lastTransitionTime":"2026-02-15T20:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.298003 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.298042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.298054 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.298067 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.298085 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:46Z","lastTransitionTime":"2026-02-15T20:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.401673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.401735 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.401755 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.401779 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.401797 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:46Z","lastTransitionTime":"2026-02-15T20:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.505080 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.505132 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.505152 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.505184 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.505203 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:46Z","lastTransitionTime":"2026-02-15T20:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.609125 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.609174 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.609183 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.609200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.609214 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:46Z","lastTransitionTime":"2026-02-15T20:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.712775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.712828 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.712845 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.712869 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.712889 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:46Z","lastTransitionTime":"2026-02-15T20:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.815528 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.815591 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.815602 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.815622 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.815646 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:46Z","lastTransitionTime":"2026-02-15T20:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.864451 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 13:06:45.492947527 +0000 UTC Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.912504 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.918644 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.918887 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.919091 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.919227 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:46 crc kubenswrapper[4735]: I0215 20:17:46.919357 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:46Z","lastTransitionTime":"2026-02-15T20:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.023085 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.023144 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.023158 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.023177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.023192 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.126872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.127287 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.127430 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.127575 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.127730 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.231702 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.231775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.231796 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.231824 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.231843 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.334923 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.334972 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.334983 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.335001 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.335014 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.438022 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.438091 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.438115 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.438144 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.438163 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.541370 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.541432 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.541452 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.541482 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.541502 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.645838 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.645906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.645923 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.645974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.645993 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.750500 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.750583 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.750608 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.750643 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.750673 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.854563 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.854620 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.854637 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.854662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.854683 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.864673 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 01:09:03.844023886 +0000 UTC Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.886454 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.886474 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.886565 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.886488 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:47 crc kubenswrapper[4735]: E0215 20:17:47.886700 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:47 crc kubenswrapper[4735]: E0215 20:17:47.886797 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:47 crc kubenswrapper[4735]: E0215 20:17:47.886931 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:47 crc kubenswrapper[4735]: E0215 20:17:47.887112 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.957361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.957417 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.957438 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.957465 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:47 crc kubenswrapper[4735]: I0215 20:17:47.957484 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:47Z","lastTransitionTime":"2026-02-15T20:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.060357 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.060427 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.060446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.060472 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.060491 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.164284 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.164381 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.164409 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.164442 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.164466 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.267775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.267825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.267844 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.267869 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.267888 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.372214 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.372283 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.372301 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.372327 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.372346 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.475480 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.475613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.475639 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.475674 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.475699 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.580116 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.580168 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.580186 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.580213 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.580233 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.683554 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.683611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.683631 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.683655 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.683673 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.789022 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.789087 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.789104 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.789133 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.789154 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.865733 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 00:57:23.801209981 +0000 UTC Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.892913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.893023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.893043 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.893068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.893088 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.996081 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.996142 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.996154 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.996183 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:48 crc kubenswrapper[4735]: I0215 20:17:48.996199 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:48Z","lastTransitionTime":"2026-02-15T20:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.099848 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.099950 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.100009 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.100042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.100065 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:49Z","lastTransitionTime":"2026-02-15T20:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.204596 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.204770 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.204789 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.204834 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.204855 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:49Z","lastTransitionTime":"2026-02-15T20:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.308835 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.308903 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.308922 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.308988 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.309014 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:49Z","lastTransitionTime":"2026-02-15T20:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.412654 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.412729 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.412748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.412784 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.412810 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:49Z","lastTransitionTime":"2026-02-15T20:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.516193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.516277 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.516290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.516313 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.516327 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:49Z","lastTransitionTime":"2026-02-15T20:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.620463 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.620528 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.620546 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.620573 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.620591 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:49Z","lastTransitionTime":"2026-02-15T20:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.724782 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.724855 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.724866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.724925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.724939 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:49Z","lastTransitionTime":"2026-02-15T20:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.828802 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.828885 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.828904 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.828934 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.828997 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:49Z","lastTransitionTime":"2026-02-15T20:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.866559 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 11:13:12.779555147 +0000 UTC Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.886083 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.886178 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.886202 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.886385 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:49 crc kubenswrapper[4735]: E0215 20:17:49.886375 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:49 crc kubenswrapper[4735]: E0215 20:17:49.886629 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:49 crc kubenswrapper[4735]: E0215 20:17:49.886910 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:49 crc kubenswrapper[4735]: E0215 20:17:49.887033 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.932990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.933076 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.933103 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.933141 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:49 crc kubenswrapper[4735]: I0215 20:17:49.933169 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:49Z","lastTransitionTime":"2026-02-15T20:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.037451 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.037518 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.037535 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.037559 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.037577 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.141437 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.141503 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.141521 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.141550 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.142251 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.247075 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.247139 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.247157 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.247184 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.247204 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.350630 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.350701 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.350720 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.350745 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.350763 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.453493 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.453563 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.453582 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.453612 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.453634 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.557351 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.557418 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.557433 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.557456 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.557472 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.661072 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.661138 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.661162 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.661188 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.661204 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.763607 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.763716 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.763742 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.763771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.763815 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.866775 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 22:33:47.486116384 +0000 UTC Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.867494 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.867536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.867558 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.867587 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.867611 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.971616 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.971687 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.971705 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.971730 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:50 crc kubenswrapper[4735]: I0215 20:17:50.971750 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:50Z","lastTransitionTime":"2026-02-15T20:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.075643 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.075730 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.075755 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.076256 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.076494 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.115932 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.116014 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.116034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.116061 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.116080 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.141153 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:51Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.148883 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.148984 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.149007 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.149036 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.149055 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.173821 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:51Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.181142 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.181199 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.181274 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.181308 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.181332 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.204666 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:51Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.210924 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.211010 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.211034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.211060 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.211083 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.233074 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:51Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.238368 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.238405 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.238422 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.238444 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.238460 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.263467 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b29648d8-57a5-4453-a2ae-0d1a863f2ae4\\\",\\\"systemUUID\\\":\\\"da2405e5-a6ff-4f93-8eee-d5a8c6845038\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:51Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.263705 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.266346 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.266407 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.266426 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.266458 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.266483 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.371178 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.371247 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.371266 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.371290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.371338 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.474699 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.474771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.474785 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.474812 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.474826 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.578130 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.578211 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.578229 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.578257 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.578278 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.682065 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.682139 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.682161 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.682193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.682217 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.785558 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.785646 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.785663 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.785688 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.785784 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.866924 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 16:29:11.199655294 +0000 UTC Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.886453 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.886492 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.886502 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.886454 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.886620 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.886788 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.886914 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:51 crc kubenswrapper[4735]: E0215 20:17:51.887178 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.888605 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.888697 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.888723 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.888753 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.888776 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.992561 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.992620 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.992637 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.992662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:51 crc kubenswrapper[4735]: I0215 20:17:51.992680 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:51Z","lastTransitionTime":"2026-02-15T20:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.097141 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.097299 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.097329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.097366 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.097390 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:52Z","lastTransitionTime":"2026-02-15T20:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.200276 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.200348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.200376 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.200407 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.200430 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:52Z","lastTransitionTime":"2026-02-15T20:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.303498 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.303551 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.303567 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.303592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.303609 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:52Z","lastTransitionTime":"2026-02-15T20:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.407172 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.407234 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.407253 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.407280 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.407303 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:52Z","lastTransitionTime":"2026-02-15T20:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.509154 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.509196 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.509205 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.509239 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.509250 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:52Z","lastTransitionTime":"2026-02-15T20:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.612621 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.612690 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.612708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.612736 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.612758 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:52Z","lastTransitionTime":"2026-02-15T20:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.716443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.716516 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.716536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.716565 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.716584 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:52Z","lastTransitionTime":"2026-02-15T20:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.820349 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.820432 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.820454 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.820490 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.820515 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:52Z","lastTransitionTime":"2026-02-15T20:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.867564 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 11:51:43.814010363 +0000 UTC Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.887924 4735 scope.go:117] "RemoveContainer" containerID="0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.910844 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba92e95-6735-4b92-b51c-4358fe0bbea1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fc54fa7126011e839c783514b4d553d25a86a4175c053e84301561b281d19fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa568691661d11a2d6c03d89986f384f64101f0a2047d4865cc4a4b04fc2483c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740e7c6d10a65b9d2bdc9cba9d431aabb9c20e9bcce1fb2f40366a8a53a4f931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:52Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.925203 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.925295 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.925314 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.925375 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.925393 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:52Z","lastTransitionTime":"2026-02-15T20:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.936542 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:52Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.963071 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"2026-02-15T20:16:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd\\\\n2026-02-15T20:16:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd to /host/opt/cni/bin/\\\\n2026-02-15T20:16:57Z [verbose] multus-daemon started\\\\n2026-02-15T20:16:57Z [verbose] Readiness Indicator file check\\\\n2026-02-15T20:17:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:52Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:52 crc kubenswrapper[4735]: I0215 20:17:52.986044 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:52Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.007566 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.028231 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.028303 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.028325 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.028356 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.028379 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.029772 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbcd727-c8da-4aa0-b937-0ee4f6c49baf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef07321fa57ad8cbabfff7062bc63516c964e96f75fbd6409dc53c63a476a228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97918334d9d8c784029961323c130538e12945a02fe644a5a44d206ea6120bf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97918334d9d8c784029961323c130538e12945a02fe644a5a44d206ea6120bf1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.062284 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.084640 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.099339 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.120146 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.132807 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.132866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.132878 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.132896 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.132930 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.141665 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.163199 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.185001 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.220737 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.236148 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.236193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.236212 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.236238 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.236259 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.239315 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.270549 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.295177 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.320472 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:53Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.339354 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.339432 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.339448 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.339470 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.340026 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.444471 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.444517 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.444535 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.444559 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.444577 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.510726 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/2.log" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.514139 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.547650 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.547692 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.547706 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.547725 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.547737 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.651604 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.651666 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.651679 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.651694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.651706 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.754487 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.754523 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.754536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.754556 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.754569 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.857439 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.857541 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.857565 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.857642 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.857669 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.868759 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 14:27:26.513577811 +0000 UTC Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.886308 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.886390 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.886314 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:53 crc kubenswrapper[4735]: E0215 20:17:53.886481 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:53 crc kubenswrapper[4735]: E0215 20:17:53.886697 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.886726 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:53 crc kubenswrapper[4735]: E0215 20:17:53.886858 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:53 crc kubenswrapper[4735]: E0215 20:17:53.886959 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.960980 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.961026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.961038 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.961054 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:53 crc kubenswrapper[4735]: I0215 20:17:53.961064 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:53Z","lastTransitionTime":"2026-02-15T20:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.066719 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.066766 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.066776 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.066794 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.066806 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.170254 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.170631 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.170642 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.170658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.170669 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.272862 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.272916 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.272931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.272971 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.272984 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.375708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.375780 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.375799 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.375824 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.375844 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.478355 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.478418 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.478433 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.478459 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.478476 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.517271 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.537323 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.553208 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.571389 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.581079 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.581144 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.581158 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.581177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.581191 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.587703 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.603610 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba92e95-6735-4b92-b51c-4358fe0bbea1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fc54fa7126011e839c783514b4d553d25a86a4175c053e84301561b281d19fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa568691661d11a2d6c03d89986f384f64101f0a2047d4865cc4a4b04fc2483c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740e7c6d10a65b9d2bdc9cba9d431aabb9c20e9bcce1fb2f40366a8a53a4f931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.619879 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.638400 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"2026-02-15T20:16:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd\\\\n2026-02-15T20:16:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd to /host/opt/cni/bin/\\\\n2026-02-15T20:16:57Z [verbose] multus-daemon started\\\\n2026-02-15T20:16:57Z [verbose] Readiness Indicator file check\\\\n2026-02-15T20:17:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.652400 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.665466 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.678401 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.683189 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.683232 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.683246 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.683269 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.683282 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.691054 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbcd727-c8da-4aa0-b937-0ee4f6c49baf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef07321fa57ad8cbabfff7062bc63516c964e96f75fbd6409dc53c63a476a228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97918334d9d8c784029961323c130538e12945a02fe644a5a44d206ea6120bf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97918334d9d8c784029961323c130538e12945a02fe644a5a44d206ea6120bf1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.708313 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.724071 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.736823 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.756650 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.767550 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.786001 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.786053 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.786070 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.786095 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.786116 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.787206 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.813855 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:54Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.869850 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 16:48:22.50194891 +0000 UTC Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.890170 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.890221 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.890237 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.890260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.890277 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.993584 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.993653 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.993668 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.993699 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:54 crc kubenswrapper[4735]: I0215 20:17:54.993722 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:54Z","lastTransitionTime":"2026-02-15T20:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.097331 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.097441 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.097455 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.097476 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.097490 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:55Z","lastTransitionTime":"2026-02-15T20:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.219719 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.219792 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.219815 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.219845 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.219867 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:55Z","lastTransitionTime":"2026-02-15T20:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.322214 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.322246 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.322255 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.322269 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.322278 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:55Z","lastTransitionTime":"2026-02-15T20:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.425419 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.425479 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.425498 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.425522 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.425540 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:55Z","lastTransitionTime":"2026-02-15T20:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.524794 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/3.log" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.526102 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/2.log" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.528437 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.528490 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.528510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.528536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.528552 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:55Z","lastTransitionTime":"2026-02-15T20:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.531333 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f" exitCode=1 Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.531385 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.531492 4735 scope.go:117] "RemoveContainer" containerID="0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.532812 4735 scope.go:117] "RemoveContainer" containerID="dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f" Feb 15 20:17:55 crc kubenswrapper[4735]: E0215 20:17:55.533181 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.553031 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbcd727-c8da-4aa0-b937-0ee4f6c49baf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef07321fa57ad8cbabfff7062bc63516c964e96f75fbd6409dc53c63a476a228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97918334d9d8c784029961323c130538e12945a02fe644a5a44d206ea6120bf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97918334d9d8c784029961323c130538e12945a02fe644a5a44d206ea6120bf1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.573127 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3afcf600-9152-49c5-bdc9-174b25f3a1a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833f3d7d5d1c2b3366738fae3c8fe79a0914cd2aa2b3b409bbe36aecfb5d0d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22794aee99e61a0ede06f5e545aa868fc61a2959f7c78b71f5c3a590ec0f028f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://805a79b943b17b50105d8002201214a7859f913460a7a3cb701e2c4b53e6eb33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.588243 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.604629 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sdcbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d52e5261-12ff-423f-ace5-22da40e004e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f28e5f0d7f78dce8a371d3ed6ad2f65531958853ced25174087674077f7554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddxhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sdcbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.622826 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43de4639-a922-4182-9377-d1f28b3aa4c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://991ce9c552216d408b90205ada67084444c2ff800250c288d7274792e69207c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz4xs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kjps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.632131 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.632201 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.632221 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.632304 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.632328 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:55Z","lastTransitionTime":"2026-02-15T20:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.645557 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.663356 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402b377399d5e20c9170a4244de1c870aee4a910d5456b4cc38bc4b42390f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.685361 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf5a527fda7caada2ffcd36a909ae3882601e0390aaf002e9d27dafe3afa397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.721772 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"755c2f64-a91b-443f-96ca-88c8cab0d656\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f5d3bbb869ce0d1b293346b9c9cb1ae364671c9f1c96338bee64f16f4cb3436\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:24Z\\\",\\\"message\\\":\\\"ice k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0215 20:17:24.031886 6277 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0215 20:17:24.032143 6277 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:54Z\\\",\\\"message\\\":\\\"0215 20:17:54.501708 6644 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:54.501757 6644 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:54.501793 6644 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:54.501845 6644 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:54.502167 6644 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:54.502428 6644 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0215 20:17:54.502441 6644 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0215 20:17:54.502479 6644 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0215 20:17:54.502493 6644 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0215 20:17:54.502528 6644 factory.go:656] Stopping watch factory\\\\nI0215 20:17:54.502550 6644 handler.go:208] Removed *v1.Node event handler 7\\\\nI0215 20:17:54.502562 6644 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0215 20:17:54.502574 6644 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nk9bx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x9xmf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.736183 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.736221 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.736239 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.736263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.736281 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:55Z","lastTransitionTime":"2026-02-15T20:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.742896 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96681837-f2a3-42d2-9653-a3997f396291\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgvvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7vfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.770023 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-15T20:16:46Z\\\",\\\"message\\\":\\\"W0215 20:16:36.293290 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0215 20:16:36.293640 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771186596 cert, and key in /tmp/serving-cert-3152889513/serving-signer.crt, /tmp/serving-cert-3152889513/serving-signer.key\\\\nI0215 20:16:36.555388 1 observer_polling.go:159] Starting file observer\\\\nW0215 20:16:36.558299 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0215 20:16:36.558689 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0215 20:16:36.562866 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3152889513/tls.crt::/tmp/serving-cert-3152889513/tls.key\\\\\\\"\\\\nF0215 20:16:46.896237 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.792525 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bca0494eaf48c0f97b4c0e437786d362ccbad8998136836004a911cabbb66654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239136fb7479e96eec337bf7b1f09b64fdd7ffa138292d69a6c15800b76ed79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.819329 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fds4z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c684f83-2a4f-465d-95f1-43e54317d164\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbd169ab1bb2b47d58bebb608a907df92dcd24ee726e92f4b5c5310c12239c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84a3896bdeebe4aa42174115b8986491933d4b9976f6e918e18887b6e4987714\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b40ea2ea20fc7fc734310a848f7a0e8592fcb89320f881412aafb54a671df76d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5705ceb3a652ea8e1806d13b8589cfc10c125e6830cd774b2d1d8ed3fbe7bbaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52fc2cc4465a51b00f0ca80cade6e35a5d8c65f7731e9a4169f7daba9d6d11fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625d444f24d80bbcf0ee266f5b6ebd2055561605dda923c0b74ef0b3abcf9b71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a78b7d51bddcf5c0f66f8cec075a0a21199c4667dea816daeb06855178281b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:17:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:17:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpdwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fds4z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.840225 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.840344 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.840368 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.840404 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.840429 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:55Z","lastTransitionTime":"2026-02-15T20:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.842230 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba92e95-6735-4b92-b51c-4358fe0bbea1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fc54fa7126011e839c783514b4d553d25a86a4175c053e84301561b281d19fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa568691661d11a2d6c03d89986f384f64101f0a2047d4865cc4a4b04fc2483c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740e7c6d10a65b9d2bdc9cba9d431aabb9c20e9bcce1fb2f40366a8a53a4f931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0e9216c75d80911d9d1db82bcc868af5d652caf3a6b6a214c6e660377316433\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-15T20:16:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.864685 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.870877 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 08:05:56.481916192 +0000 UTC Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.886274 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.886333 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:55 crc kubenswrapper[4735]: E0215 20:17:55.886415 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.886345 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.886333 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:55 crc kubenswrapper[4735]: E0215 20:17:55.886521 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:55 crc kubenswrapper[4735]: E0215 20:17:55.886669 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.886823 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wfjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a20277cf-39a2-453f-ac6d-fb28346b3358\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-15T20:17:42Z\\\",\\\"message\\\":\\\"2026-02-15T20:16:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd\\\\n2026-02-15T20:16:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3f1200d6-29ac-42dd-8bae-3de5eae5f0cd to /host/opt/cni/bin/\\\\n2026-02-15T20:16:57Z [verbose] multus-daemon started\\\\n2026-02-15T20:16:57Z [verbose] Readiness Indicator file check\\\\n2026-02-15T20:17:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-15T20:16:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmcnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wfjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: E0215 20:17:55.887006 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.902730 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9ls74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1df5b8f-f6f3-45a6-99d2-088da4f8981b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f00880783a8140193b812c0121c97652efc684e44ce826bdb07553b8d3ce23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6pvb8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:16:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9ls74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.923793 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92a41601-6d64-46c3-a9ea-2fe0427d884b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-15T20:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd0dea5539b07a09b671361b375fea6003f2fe54369b8ee3ddd4be543a64d91d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://755d5c97d180794d9757f3dd811b452562bb96317ff56e60fbc7b2af8dd98b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-15T20:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vbqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-15T20:17:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q6fbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-15T20:17:55Z is after 2025-08-24T17:21:41Z" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.944040 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.944332 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.944544 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.944760 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:55 crc kubenswrapper[4735]: I0215 20:17:55.944992 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:55Z","lastTransitionTime":"2026-02-15T20:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.048169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.048217 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.048228 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.048245 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.048257 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.150546 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.150582 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.150592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.150607 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.150618 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.253347 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.253409 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.253423 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.253446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.253460 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.356419 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.356479 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.356497 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.356525 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.356547 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.459765 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.459825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.459842 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.459868 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.460109 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.537776 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/3.log" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.543353 4735 scope.go:117] "RemoveContainer" containerID="dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f" Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.543498 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.563397 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.563476 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.563497 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.563529 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.563555 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.619292 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-wfjmx" podStartSLOduration=62.619259026 podStartE2EDuration="1m2.619259026s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.603430616 +0000 UTC m=+84.469446299" watchObservedRunningTime="2026-02-15 20:17:56.619259026 +0000 UTC m=+84.485274679" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.619751 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-9ls74" podStartSLOduration=63.619737161 podStartE2EDuration="1m3.619737161s" podCreationTimestamp="2026-02-15 20:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.617025993 +0000 UTC m=+84.483041606" watchObservedRunningTime="2026-02-15 20:17:56.619737161 +0000 UTC m=+84.485752814" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.642430 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q6fbm" podStartSLOduration=62.642408936 podStartE2EDuration="1m2.642408936s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.638876844 +0000 UTC m=+84.504892527" watchObservedRunningTime="2026-02-15 20:17:56.642408936 +0000 UTC m=+84.508424559" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.666869 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.667009 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.667038 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.667079 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.667110 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.699250 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=29.699217852 podStartE2EDuration="29.699217852s" podCreationTimestamp="2026-02-15 20:17:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.671590166 +0000 UTC m=+84.537605849" watchObservedRunningTime="2026-02-15 20:17:56.699217852 +0000 UTC m=+84.565233495" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.699727 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.699720216 podStartE2EDuration="10.699720216s" podCreationTimestamp="2026-02-15 20:17:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.692524461 +0000 UTC m=+84.558540134" watchObservedRunningTime="2026-02-15 20:17:56.699720216 +0000 UTC m=+84.565735849" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.735017 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=62.73499437 podStartE2EDuration="1m2.73499437s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.717179773 +0000 UTC m=+84.583195426" watchObservedRunningTime="2026-02-15 20:17:56.73499437 +0000 UTC m=+84.601009983" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.761328 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-sdcbj" podStartSLOduration=63.761308889 podStartE2EDuration="1m3.761308889s" podCreationTimestamp="2026-02-15 20:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.748247137 +0000 UTC m=+84.614262760" watchObservedRunningTime="2026-02-15 20:17:56.761308889 +0000 UTC m=+84.627324502" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.769923 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.769982 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.769994 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.770011 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.770024 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.776834 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podStartSLOduration=62.77681458 podStartE2EDuration="1m2.77681458s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.761692149 +0000 UTC m=+84.627707772" watchObservedRunningTime="2026-02-15 20:17:56.77681458 +0000 UTC m=+84.642830213" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.837255 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.837422 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.837395454 +0000 UTC m=+148.703411077 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.837460 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.837526 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.837567 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.837619 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.8376081 +0000 UTC m=+148.703623733 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.837789 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.837936 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.837906268 +0000 UTC m=+148.703922081 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.871078 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 19:14:27.458920567 +0000 UTC Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.873324 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.873360 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.873370 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.873392 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.873403 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.892639 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-fds4z" podStartSLOduration=62.892622145 podStartE2EDuration="1m2.892622145s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.892281086 +0000 UTC m=+84.758296719" watchObservedRunningTime="2026-02-15 20:17:56.892622145 +0000 UTC m=+84.758637768" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.909727 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=65.909708261 podStartE2EDuration="1m5.909708261s" podCreationTimestamp="2026-02-15 20:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:17:56.908613151 +0000 UTC m=+84.774628784" watchObservedRunningTime="2026-02-15 20:17:56.909708261 +0000 UTC m=+84.775723884" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.937976 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.938044 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.938167 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.938194 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.938208 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.938258 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.938243403 +0000 UTC m=+148.804259026 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.938168 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.938305 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.938316 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:56 crc kubenswrapper[4735]: E0215 20:17:56.938358 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.938346656 +0000 UTC m=+148.804362279 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.976001 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.976057 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.976070 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.976092 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:56 crc kubenswrapper[4735]: I0215 20:17:56.976105 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:56Z","lastTransitionTime":"2026-02-15T20:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.078649 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.078694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.078706 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.078724 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.078738 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:57Z","lastTransitionTime":"2026-02-15T20:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.181493 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.181546 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.181557 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.181574 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.181587 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:57Z","lastTransitionTime":"2026-02-15T20:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.285504 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.285572 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.285588 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.285614 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.285633 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:57Z","lastTransitionTime":"2026-02-15T20:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.388625 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.388686 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.388703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.388727 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.388745 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:57Z","lastTransitionTime":"2026-02-15T20:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.492467 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.492543 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.492556 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.492576 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.492589 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:57Z","lastTransitionTime":"2026-02-15T20:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.596138 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.596221 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.596245 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.596286 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.596317 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:57Z","lastTransitionTime":"2026-02-15T20:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.699974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.700056 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.700082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.700116 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.700142 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:57Z","lastTransitionTime":"2026-02-15T20:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.804119 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.804255 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.804282 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.804320 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.804345 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:57Z","lastTransitionTime":"2026-02-15T20:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.871924 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 03:24:16.101349598 +0000 UTC Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.886270 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.886269 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.886333 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.886495 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:57 crc kubenswrapper[4735]: E0215 20:17:57.886699 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:57 crc kubenswrapper[4735]: E0215 20:17:57.887007 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:57 crc kubenswrapper[4735]: E0215 20:17:57.887203 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:57 crc kubenswrapper[4735]: E0215 20:17:57.887306 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.908139 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.908195 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.908214 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.908240 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:57 crc kubenswrapper[4735]: I0215 20:17:57.908258 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:57Z","lastTransitionTime":"2026-02-15T20:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.012185 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.012268 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.012287 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.012326 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.012351 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.115883 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.116032 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.116055 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.116081 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.116098 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.220114 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.220177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.220195 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.220219 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.220240 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.324262 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.324350 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.324373 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.324404 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.324426 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.428045 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.428118 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.428136 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.428170 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.428190 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.531927 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.532036 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.532060 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.532101 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.532124 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.636645 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.636717 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.636738 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.636764 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.636787 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.740528 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.740606 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.740625 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.740658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.740676 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.843991 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.844062 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.844082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.844112 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.844135 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.873036 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 12:06:09.760888924 +0000 UTC Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.947589 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.948180 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.948385 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.948692 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:58 crc kubenswrapper[4735]: I0215 20:17:58.948874 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:58Z","lastTransitionTime":"2026-02-15T20:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.052011 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.052056 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.052068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.052089 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.052102 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.155673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.155755 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.155775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.155804 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.155825 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.260034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.260103 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.260120 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.260144 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.260162 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.363865 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.364146 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.364195 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.364224 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.364543 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.468308 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.468384 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.468410 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.468441 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.468463 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.571237 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.571309 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.571329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.571353 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.571370 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.674344 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.674415 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.674433 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.674459 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.674477 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.777505 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.777596 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.777621 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.777654 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.777685 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.873635 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 05:58:22.738499625 +0000 UTC Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.881423 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.881498 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.881519 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.881549 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.881569 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.886782 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.886815 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.886846 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.886875 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:17:59 crc kubenswrapper[4735]: E0215 20:17:59.887000 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:17:59 crc kubenswrapper[4735]: E0215 20:17:59.887118 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:17:59 crc kubenswrapper[4735]: E0215 20:17:59.887233 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:17:59 crc kubenswrapper[4735]: E0215 20:17:59.887334 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.984122 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.984189 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.984208 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.984235 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:17:59 crc kubenswrapper[4735]: I0215 20:17:59.984254 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:17:59Z","lastTransitionTime":"2026-02-15T20:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.087899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.087998 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.088019 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.088045 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.088065 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:00Z","lastTransitionTime":"2026-02-15T20:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.191130 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.191238 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.191257 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.191285 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.191304 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:00Z","lastTransitionTime":"2026-02-15T20:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.294751 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.294806 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.294827 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.294855 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.294874 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:00Z","lastTransitionTime":"2026-02-15T20:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.398510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.398563 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.398581 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.398607 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.398625 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:00Z","lastTransitionTime":"2026-02-15T20:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.501334 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.501408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.501436 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.501467 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.501493 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:00Z","lastTransitionTime":"2026-02-15T20:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.605316 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.606090 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.606405 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.606578 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.606735 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:00Z","lastTransitionTime":"2026-02-15T20:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.711274 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.711340 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.711358 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.711383 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.711401 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:00Z","lastTransitionTime":"2026-02-15T20:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.814491 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.814544 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.814561 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.814585 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.814603 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:00Z","lastTransitionTime":"2026-02-15T20:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.874010 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 19:32:30.813035478 +0000 UTC Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.917377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.917402 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.917411 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.917425 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:00 crc kubenswrapper[4735]: I0215 20:18:00.917433 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:00Z","lastTransitionTime":"2026-02-15T20:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.019823 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.019851 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.019859 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.019872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.019881 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:01Z","lastTransitionTime":"2026-02-15T20:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.122407 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.122474 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.122493 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.122520 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.122539 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:01Z","lastTransitionTime":"2026-02-15T20:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.226992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.227088 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.227115 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.227189 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.227216 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:01Z","lastTransitionTime":"2026-02-15T20:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.330457 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.330520 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.330538 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.330565 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.330586 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:01Z","lastTransitionTime":"2026-02-15T20:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.401348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.401445 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.401468 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.401841 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.402127 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-15T20:18:01Z","lastTransitionTime":"2026-02-15T20:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.476585 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl"] Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.477192 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.483070 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.484119 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.485243 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.487491 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.598146 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8acc2747-b5d4-4f53-960b-157ebf33d50c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.598244 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8acc2747-b5d4-4f53-960b-157ebf33d50c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.598299 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8acc2747-b5d4-4f53-960b-157ebf33d50c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.598487 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8acc2747-b5d4-4f53-960b-157ebf33d50c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.598533 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8acc2747-b5d4-4f53-960b-157ebf33d50c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.699310 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8acc2747-b5d4-4f53-960b-157ebf33d50c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.699410 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8acc2747-b5d4-4f53-960b-157ebf33d50c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.699425 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8acc2747-b5d4-4f53-960b-157ebf33d50c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.699581 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8acc2747-b5d4-4f53-960b-157ebf33d50c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.699648 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8acc2747-b5d4-4f53-960b-157ebf33d50c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.699698 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8acc2747-b5d4-4f53-960b-157ebf33d50c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.699718 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8acc2747-b5d4-4f53-960b-157ebf33d50c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.701071 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8acc2747-b5d4-4f53-960b-157ebf33d50c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.710347 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8acc2747-b5d4-4f53-960b-157ebf33d50c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.740279 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8acc2747-b5d4-4f53-960b-157ebf33d50c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9bhwl\" (UID: \"8acc2747-b5d4-4f53-960b-157ebf33d50c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.806653 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.874994 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 20:42:57.847362914 +0000 UTC Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.875068 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.885942 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:01 crc kubenswrapper[4735]: E0215 20:18:01.886183 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.886273 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:01 crc kubenswrapper[4735]: E0215 20:18:01.886357 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.886431 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:01 crc kubenswrapper[4735]: E0215 20:18:01.886516 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.887342 4735 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 15 20:18:01 crc kubenswrapper[4735]: I0215 20:18:01.887702 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:01 crc kubenswrapper[4735]: E0215 20:18:01.887833 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:02 crc kubenswrapper[4735]: I0215 20:18:02.568833 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" event={"ID":"8acc2747-b5d4-4f53-960b-157ebf33d50c","Type":"ContainerStarted","Data":"cab0b3b3e3fb944796a1f78930b29f5e209ae4a977ea666ba621584019f32a90"} Feb 15 20:18:02 crc kubenswrapper[4735]: I0215 20:18:02.568890 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" event={"ID":"8acc2747-b5d4-4f53-960b-157ebf33d50c","Type":"ContainerStarted","Data":"76f52bbf83b646e6dd49980a949851f67ef4c596c185265f1c6be8903f6ab07d"} Feb 15 20:18:03 crc kubenswrapper[4735]: I0215 20:18:03.886207 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:03 crc kubenswrapper[4735]: I0215 20:18:03.886205 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:03 crc kubenswrapper[4735]: I0215 20:18:03.886336 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:03 crc kubenswrapper[4735]: I0215 20:18:03.886383 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:03 crc kubenswrapper[4735]: E0215 20:18:03.886559 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:03 crc kubenswrapper[4735]: E0215 20:18:03.886942 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:03 crc kubenswrapper[4735]: E0215 20:18:03.887266 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:03 crc kubenswrapper[4735]: E0215 20:18:03.887386 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:05 crc kubenswrapper[4735]: I0215 20:18:05.886067 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:05 crc kubenswrapper[4735]: E0215 20:18:05.886887 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:05 crc kubenswrapper[4735]: I0215 20:18:05.886261 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:05 crc kubenswrapper[4735]: E0215 20:18:05.887232 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:05 crc kubenswrapper[4735]: I0215 20:18:05.886313 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:05 crc kubenswrapper[4735]: E0215 20:18:05.887463 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:05 crc kubenswrapper[4735]: I0215 20:18:05.886189 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:05 crc kubenswrapper[4735]: E0215 20:18:05.887696 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:07 crc kubenswrapper[4735]: I0215 20:18:07.886546 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:07 crc kubenswrapper[4735]: I0215 20:18:07.886635 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:07 crc kubenswrapper[4735]: I0215 20:18:07.886586 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:07 crc kubenswrapper[4735]: E0215 20:18:07.886826 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:07 crc kubenswrapper[4735]: E0215 20:18:07.886939 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:07 crc kubenswrapper[4735]: I0215 20:18:07.887002 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:07 crc kubenswrapper[4735]: E0215 20:18:07.887105 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:07 crc kubenswrapper[4735]: E0215 20:18:07.887353 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:09 crc kubenswrapper[4735]: I0215 20:18:09.886401 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:09 crc kubenswrapper[4735]: I0215 20:18:09.886552 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:09 crc kubenswrapper[4735]: E0215 20:18:09.887231 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:09 crc kubenswrapper[4735]: I0215 20:18:09.886592 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:09 crc kubenswrapper[4735]: E0215 20:18:09.887289 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:09 crc kubenswrapper[4735]: I0215 20:18:09.886657 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:09 crc kubenswrapper[4735]: E0215 20:18:09.887487 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:09 crc kubenswrapper[4735]: E0215 20:18:09.887522 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:10 crc kubenswrapper[4735]: I0215 20:18:10.889681 4735 scope.go:117] "RemoveContainer" containerID="dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f" Feb 15 20:18:10 crc kubenswrapper[4735]: E0215 20:18:10.890143 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" Feb 15 20:18:11 crc kubenswrapper[4735]: I0215 20:18:11.886352 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:11 crc kubenswrapper[4735]: I0215 20:18:11.886420 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:11 crc kubenswrapper[4735]: I0215 20:18:11.886987 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:11 crc kubenswrapper[4735]: E0215 20:18:11.887156 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:11 crc kubenswrapper[4735]: I0215 20:18:11.887215 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:11 crc kubenswrapper[4735]: E0215 20:18:11.887578 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:11 crc kubenswrapper[4735]: E0215 20:18:11.887338 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:11 crc kubenswrapper[4735]: E0215 20:18:11.888026 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:13 crc kubenswrapper[4735]: I0215 20:18:13.547559 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:13 crc kubenswrapper[4735]: E0215 20:18:13.547766 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:18:13 crc kubenswrapper[4735]: E0215 20:18:13.547913 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs podName:96681837-f2a3-42d2-9653-a3997f396291 nodeName:}" failed. No retries permitted until 2026-02-15 20:19:17.547840302 +0000 UTC m=+165.413855955 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs") pod "network-metrics-daemon-7vfs4" (UID: "96681837-f2a3-42d2-9653-a3997f396291") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 15 20:18:13 crc kubenswrapper[4735]: I0215 20:18:13.886334 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:13 crc kubenswrapper[4735]: I0215 20:18:13.886451 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:13 crc kubenswrapper[4735]: I0215 20:18:13.886334 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:13 crc kubenswrapper[4735]: E0215 20:18:13.886559 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:13 crc kubenswrapper[4735]: E0215 20:18:13.886721 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:13 crc kubenswrapper[4735]: I0215 20:18:13.887056 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:13 crc kubenswrapper[4735]: E0215 20:18:13.887046 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:13 crc kubenswrapper[4735]: E0215 20:18:13.887203 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:15 crc kubenswrapper[4735]: I0215 20:18:15.885934 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:15 crc kubenswrapper[4735]: E0215 20:18:15.886244 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:15 crc kubenswrapper[4735]: I0215 20:18:15.886383 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:15 crc kubenswrapper[4735]: E0215 20:18:15.886574 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:15 crc kubenswrapper[4735]: I0215 20:18:15.886854 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:15 crc kubenswrapper[4735]: I0215 20:18:15.887032 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:15 crc kubenswrapper[4735]: E0215 20:18:15.887039 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:15 crc kubenswrapper[4735]: E0215 20:18:15.887423 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:17 crc kubenswrapper[4735]: I0215 20:18:17.886481 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:17 crc kubenswrapper[4735]: I0215 20:18:17.886534 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:17 crc kubenswrapper[4735]: I0215 20:18:17.886534 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:17 crc kubenswrapper[4735]: I0215 20:18:17.886541 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:17 crc kubenswrapper[4735]: E0215 20:18:17.886669 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:17 crc kubenswrapper[4735]: E0215 20:18:17.886841 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:17 crc kubenswrapper[4735]: E0215 20:18:17.887004 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:17 crc kubenswrapper[4735]: E0215 20:18:17.887083 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:19 crc kubenswrapper[4735]: I0215 20:18:19.886239 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:19 crc kubenswrapper[4735]: I0215 20:18:19.886354 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:19 crc kubenswrapper[4735]: E0215 20:18:19.886435 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:19 crc kubenswrapper[4735]: I0215 20:18:19.886500 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:19 crc kubenswrapper[4735]: I0215 20:18:19.886507 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:19 crc kubenswrapper[4735]: E0215 20:18:19.886621 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:19 crc kubenswrapper[4735]: E0215 20:18:19.886782 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:19 crc kubenswrapper[4735]: E0215 20:18:19.886929 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:19 crc kubenswrapper[4735]: I0215 20:18:19.911639 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bhwl" podStartSLOduration=85.911610726 podStartE2EDuration="1m25.911610726s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:02.589145995 +0000 UTC m=+90.455161618" watchObservedRunningTime="2026-02-15 20:18:19.911610726 +0000 UTC m=+107.777626389" Feb 15 20:18:19 crc kubenswrapper[4735]: I0215 20:18:19.912892 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 15 20:18:21 crc kubenswrapper[4735]: I0215 20:18:21.886490 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:21 crc kubenswrapper[4735]: I0215 20:18:21.886518 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:21 crc kubenswrapper[4735]: I0215 20:18:21.886527 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:21 crc kubenswrapper[4735]: I0215 20:18:21.886592 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:21 crc kubenswrapper[4735]: E0215 20:18:21.888247 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:21 crc kubenswrapper[4735]: E0215 20:18:21.888365 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:21 crc kubenswrapper[4735]: E0215 20:18:21.888446 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:21 crc kubenswrapper[4735]: E0215 20:18:21.888541 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:23 crc kubenswrapper[4735]: I0215 20:18:23.886594 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:23 crc kubenswrapper[4735]: I0215 20:18:23.886635 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:23 crc kubenswrapper[4735]: I0215 20:18:23.886650 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:23 crc kubenswrapper[4735]: I0215 20:18:23.886912 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:23 crc kubenswrapper[4735]: E0215 20:18:23.887817 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:23 crc kubenswrapper[4735]: E0215 20:18:23.887847 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:23 crc kubenswrapper[4735]: E0215 20:18:23.888936 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:23 crc kubenswrapper[4735]: E0215 20:18:23.889208 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:24 crc kubenswrapper[4735]: I0215 20:18:24.887746 4735 scope.go:117] "RemoveContainer" containerID="dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f" Feb 15 20:18:24 crc kubenswrapper[4735]: E0215 20:18:24.888056 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x9xmf_openshift-ovn-kubernetes(755c2f64-a91b-443f-96ca-88c8cab0d656)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" Feb 15 20:18:25 crc kubenswrapper[4735]: I0215 20:18:25.885840 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:25 crc kubenswrapper[4735]: I0215 20:18:25.885840 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:25 crc kubenswrapper[4735]: E0215 20:18:25.886505 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:25 crc kubenswrapper[4735]: I0215 20:18:25.885914 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:25 crc kubenswrapper[4735]: E0215 20:18:25.886732 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:25 crc kubenswrapper[4735]: E0215 20:18:25.886591 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:25 crc kubenswrapper[4735]: I0215 20:18:25.887072 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:25 crc kubenswrapper[4735]: E0215 20:18:25.887175 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:27 crc kubenswrapper[4735]: I0215 20:18:27.886779 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:27 crc kubenswrapper[4735]: I0215 20:18:27.886799 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:27 crc kubenswrapper[4735]: I0215 20:18:27.887044 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:27 crc kubenswrapper[4735]: E0215 20:18:27.887277 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:27 crc kubenswrapper[4735]: I0215 20:18:27.887322 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:27 crc kubenswrapper[4735]: E0215 20:18:27.887547 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:27 crc kubenswrapper[4735]: E0215 20:18:27.887682 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:27 crc kubenswrapper[4735]: E0215 20:18:27.887795 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:28 crc kubenswrapper[4735]: I0215 20:18:28.682515 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/1.log" Feb 15 20:18:28 crc kubenswrapper[4735]: I0215 20:18:28.683978 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/0.log" Feb 15 20:18:28 crc kubenswrapper[4735]: I0215 20:18:28.684193 4735 generic.go:334] "Generic (PLEG): container finished" podID="a20277cf-39a2-453f-ac6d-fb28346b3358" containerID="17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889" exitCode=1 Feb 15 20:18:28 crc kubenswrapper[4735]: I0215 20:18:28.684354 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wfjmx" event={"ID":"a20277cf-39a2-453f-ac6d-fb28346b3358","Type":"ContainerDied","Data":"17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889"} Feb 15 20:18:28 crc kubenswrapper[4735]: I0215 20:18:28.684492 4735 scope.go:117] "RemoveContainer" containerID="4acde5289157c13dca4251c06c7b249e3427144d0179faed2a2bd76408f5bef9" Feb 15 20:18:28 crc kubenswrapper[4735]: I0215 20:18:28.685319 4735 scope.go:117] "RemoveContainer" containerID="17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889" Feb 15 20:18:28 crc kubenswrapper[4735]: E0215 20:18:28.685688 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-wfjmx_openshift-multus(a20277cf-39a2-453f-ac6d-fb28346b3358)\"" pod="openshift-multus/multus-wfjmx" podUID="a20277cf-39a2-453f-ac6d-fb28346b3358" Feb 15 20:18:28 crc kubenswrapper[4735]: I0215 20:18:28.729416 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=9.729385793 podStartE2EDuration="9.729385793s" podCreationTimestamp="2026-02-15 20:18:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:22.939383963 +0000 UTC m=+110.805399636" watchObservedRunningTime="2026-02-15 20:18:28.729385793 +0000 UTC m=+116.595401446" Feb 15 20:18:29 crc kubenswrapper[4735]: I0215 20:18:29.690277 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/1.log" Feb 15 20:18:29 crc kubenswrapper[4735]: I0215 20:18:29.885887 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:29 crc kubenswrapper[4735]: I0215 20:18:29.886033 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:29 crc kubenswrapper[4735]: I0215 20:18:29.886073 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:29 crc kubenswrapper[4735]: E0215 20:18:29.886159 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:29 crc kubenswrapper[4735]: E0215 20:18:29.886280 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:29 crc kubenswrapper[4735]: I0215 20:18:29.886396 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:29 crc kubenswrapper[4735]: E0215 20:18:29.886397 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:29 crc kubenswrapper[4735]: E0215 20:18:29.886513 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:31 crc kubenswrapper[4735]: I0215 20:18:31.886417 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:31 crc kubenswrapper[4735]: I0215 20:18:31.886456 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:31 crc kubenswrapper[4735]: I0215 20:18:31.886477 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:31 crc kubenswrapper[4735]: I0215 20:18:31.886554 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:31 crc kubenswrapper[4735]: E0215 20:18:31.887461 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:31 crc kubenswrapper[4735]: E0215 20:18:31.887827 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:31 crc kubenswrapper[4735]: E0215 20:18:31.888010 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:31 crc kubenswrapper[4735]: E0215 20:18:31.888138 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:32 crc kubenswrapper[4735]: E0215 20:18:32.868095 4735 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 15 20:18:33 crc kubenswrapper[4735]: E0215 20:18:33.049648 4735 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 15 20:18:33 crc kubenswrapper[4735]: I0215 20:18:33.886596 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:33 crc kubenswrapper[4735]: I0215 20:18:33.886631 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:33 crc kubenswrapper[4735]: I0215 20:18:33.886704 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:33 crc kubenswrapper[4735]: I0215 20:18:33.886791 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:33 crc kubenswrapper[4735]: E0215 20:18:33.886900 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:33 crc kubenswrapper[4735]: E0215 20:18:33.887083 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:33 crc kubenswrapper[4735]: E0215 20:18:33.887251 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:33 crc kubenswrapper[4735]: E0215 20:18:33.887377 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:35 crc kubenswrapper[4735]: I0215 20:18:35.886665 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:35 crc kubenswrapper[4735]: I0215 20:18:35.886738 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:35 crc kubenswrapper[4735]: I0215 20:18:35.886702 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:35 crc kubenswrapper[4735]: I0215 20:18:35.886702 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:35 crc kubenswrapper[4735]: E0215 20:18:35.887018 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:35 crc kubenswrapper[4735]: E0215 20:18:35.887165 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:35 crc kubenswrapper[4735]: E0215 20:18:35.887268 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:35 crc kubenswrapper[4735]: E0215 20:18:35.887468 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:36 crc kubenswrapper[4735]: I0215 20:18:36.888434 4735 scope.go:117] "RemoveContainer" containerID="dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f" Feb 15 20:18:37 crc kubenswrapper[4735]: I0215 20:18:37.724284 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/3.log" Feb 15 20:18:37 crc kubenswrapper[4735]: I0215 20:18:37.727375 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerStarted","Data":"94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc"} Feb 15 20:18:37 crc kubenswrapper[4735]: I0215 20:18:37.727880 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:18:37 crc kubenswrapper[4735]: I0215 20:18:37.818881 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podStartSLOduration=103.81885399 podStartE2EDuration="1m43.81885399s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:37.767686873 +0000 UTC m=+125.633702516" watchObservedRunningTime="2026-02-15 20:18:37.81885399 +0000 UTC m=+125.684869623" Feb 15 20:18:37 crc kubenswrapper[4735]: I0215 20:18:37.820445 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7vfs4"] Feb 15 20:18:37 crc kubenswrapper[4735]: I0215 20:18:37.820667 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:37 crc kubenswrapper[4735]: E0215 20:18:37.820865 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:37 crc kubenswrapper[4735]: I0215 20:18:37.885703 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:37 crc kubenswrapper[4735]: I0215 20:18:37.885771 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:37 crc kubenswrapper[4735]: E0215 20:18:37.885854 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:37 crc kubenswrapper[4735]: E0215 20:18:37.885937 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:37 crc kubenswrapper[4735]: I0215 20:18:37.886180 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:37 crc kubenswrapper[4735]: E0215 20:18:37.886370 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:38 crc kubenswrapper[4735]: E0215 20:18:38.051667 4735 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 15 20:18:39 crc kubenswrapper[4735]: I0215 20:18:39.886352 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:39 crc kubenswrapper[4735]: I0215 20:18:39.886413 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:39 crc kubenswrapper[4735]: I0215 20:18:39.886384 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:39 crc kubenswrapper[4735]: I0215 20:18:39.886384 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:39 crc kubenswrapper[4735]: E0215 20:18:39.886619 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:39 crc kubenswrapper[4735]: E0215 20:18:39.886764 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:39 crc kubenswrapper[4735]: E0215 20:18:39.886884 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:39 crc kubenswrapper[4735]: E0215 20:18:39.887178 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:41 crc kubenswrapper[4735]: I0215 20:18:41.885882 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:41 crc kubenswrapper[4735]: I0215 20:18:41.885934 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:41 crc kubenswrapper[4735]: I0215 20:18:41.886093 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:41 crc kubenswrapper[4735]: E0215 20:18:41.886151 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:41 crc kubenswrapper[4735]: E0215 20:18:41.886350 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:41 crc kubenswrapper[4735]: E0215 20:18:41.886529 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:41 crc kubenswrapper[4735]: I0215 20:18:41.886878 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:41 crc kubenswrapper[4735]: E0215 20:18:41.887067 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:43 crc kubenswrapper[4735]: E0215 20:18:43.053939 4735 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 15 20:18:43 crc kubenswrapper[4735]: I0215 20:18:43.886793 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:43 crc kubenswrapper[4735]: I0215 20:18:43.886844 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:43 crc kubenswrapper[4735]: I0215 20:18:43.886908 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:43 crc kubenswrapper[4735]: I0215 20:18:43.886935 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:43 crc kubenswrapper[4735]: E0215 20:18:43.887293 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:43 crc kubenswrapper[4735]: E0215 20:18:43.887411 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:43 crc kubenswrapper[4735]: I0215 20:18:43.887428 4735 scope.go:117] "RemoveContainer" containerID="17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889" Feb 15 20:18:43 crc kubenswrapper[4735]: E0215 20:18:43.887542 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:43 crc kubenswrapper[4735]: E0215 20:18:43.887622 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:44 crc kubenswrapper[4735]: I0215 20:18:44.756438 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/1.log" Feb 15 20:18:44 crc kubenswrapper[4735]: I0215 20:18:44.756836 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wfjmx" event={"ID":"a20277cf-39a2-453f-ac6d-fb28346b3358","Type":"ContainerStarted","Data":"987c4a7181b247e73c933f1f073f70911e9100019abf6fd2a37c154a601e5c62"} Feb 15 20:18:45 crc kubenswrapper[4735]: I0215 20:18:45.886756 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:45 crc kubenswrapper[4735]: I0215 20:18:45.886809 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:45 crc kubenswrapper[4735]: I0215 20:18:45.886780 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:45 crc kubenswrapper[4735]: I0215 20:18:45.886756 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:45 crc kubenswrapper[4735]: E0215 20:18:45.887054 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:45 crc kubenswrapper[4735]: E0215 20:18:45.887119 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:45 crc kubenswrapper[4735]: E0215 20:18:45.887372 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:45 crc kubenswrapper[4735]: E0215 20:18:45.887424 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:47 crc kubenswrapper[4735]: I0215 20:18:47.885902 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:47 crc kubenswrapper[4735]: I0215 20:18:47.886085 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:47 crc kubenswrapper[4735]: E0215 20:18:47.886186 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 15 20:18:47 crc kubenswrapper[4735]: I0215 20:18:47.886203 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:47 crc kubenswrapper[4735]: I0215 20:18:47.886307 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:47 crc kubenswrapper[4735]: E0215 20:18:47.886365 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7vfs4" podUID="96681837-f2a3-42d2-9653-a3997f396291" Feb 15 20:18:47 crc kubenswrapper[4735]: E0215 20:18:47.886500 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 15 20:18:47 crc kubenswrapper[4735]: E0215 20:18:47.886616 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.885790 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.885901 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.885902 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.886527 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.890678 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.892616 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.892681 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.892936 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.892983 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 15 20:18:49 crc kubenswrapper[4735]: I0215 20:18:49.897174 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.459434 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.514617 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r89v"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.515410 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.524113 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.524262 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.524422 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.524114 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.524903 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.525252 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.525369 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.525641 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.526122 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.526566 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.526899 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.527644 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.530234 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rjhgs"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.531077 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.533154 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jbnv7"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.534016 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: W0215 20:18:52.537444 4735 reflector.go:561] object-"openshift-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Feb 15 20:18:52 crc kubenswrapper[4735]: E0215 20:18:52.537528 4735 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 15 20:18:52 crc kubenswrapper[4735]: W0215 20:18:52.541779 4735 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Feb 15 20:18:52 crc kubenswrapper[4735]: W0215 20:18:52.541828 4735 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-tls": failed to list *v1.Secret: secrets "machine-api-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Feb 15 20:18:52 crc kubenswrapper[4735]: E0215 20:18:52.541852 4735 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 15 20:18:52 crc kubenswrapper[4735]: E0215 20:18:52.541897 4735 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 15 20:18:52 crc kubenswrapper[4735]: W0215 20:18:52.542040 4735 reflector.go:561] object-"openshift-machine-api"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Feb 15 20:18:52 crc kubenswrapper[4735]: E0215 20:18:52.542079 4735 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.542213 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.543034 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: W0215 20:18:52.544906 4735 reflector.go:561] object-"openshift-machine-api"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Feb 15 20:18:52 crc kubenswrapper[4735]: E0215 20:18:52.545063 4735 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.545124 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8nz76"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.545672 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.546213 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.546700 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.550166 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.550414 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.552161 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.552337 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.552490 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.552776 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.552987 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.553069 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.553127 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.553587 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.554006 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.554873 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.554919 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-dir\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.554963 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.554989 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555009 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555030 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555051 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b885v\" (UniqueName: \"kubernetes.io/projected/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-kube-api-access-b885v\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555088 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555109 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555131 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555153 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555195 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555213 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-policies\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.555233 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.569516 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.570687 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.570847 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.571474 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.573981 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.579756 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.580055 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.581980 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.586645 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.586789 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.586660 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.588573 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.599203 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.600177 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.600382 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.600907 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.602012 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.604422 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.607395 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.617094 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jfpmq"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.617742 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.621627 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.621878 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.622079 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.622208 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.624758 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.625007 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.627300 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.627966 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.628044 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.629776 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.630099 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.630139 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.630289 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.630409 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.630517 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.630870 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.635128 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.645746 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-k64kb"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.647486 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.649459 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.649905 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-x4gh2"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.651235 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.651632 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656196 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a13a950e-dc37-4878-afdf-623a7c7d01b1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkgwz\" (UID: \"a13a950e-dc37-4878-afdf-623a7c7d01b1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656245 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-dir\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656269 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656288 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656304 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130674aa-a02b-4d35-bf24-0c74de51fe4d-config\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656322 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpkt9\" (UniqueName: \"kubernetes.io/projected/3fb3271e-4d33-4ef0-b269-ca69f698fb66-kube-api-access-fpkt9\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656344 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656361 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656377 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656393 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b885v\" (UniqueName: \"kubernetes.io/projected/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-kube-api-access-b885v\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656409 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a13a950e-dc37-4878-afdf-623a7c7d01b1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkgwz\" (UID: \"a13a950e-dc37-4878-afdf-623a7c7d01b1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656437 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656456 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656475 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/130674aa-a02b-4d35-bf24-0c74de51fe4d-service-ca-bundle\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656516 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-images\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656532 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f00c55b0-60e2-4d35-a250-706ec8d9de8a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tr8hd\" (UID: \"f00c55b0-60e2-4d35-a250-706ec8d9de8a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656550 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7d0ff181-18dc-47c5-b173-d5197b923176-auth-proxy-config\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656575 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7d0ff181-18dc-47c5-b173-d5197b923176-machine-approver-tls\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656593 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d3be498a-3362-4c71-85ac-e53b4889eec5-audit-dir\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656610 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x445g\" (UniqueName: \"kubernetes.io/projected/d3be498a-3362-4c71-85ac-e53b4889eec5-kube-api-access-x445g\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656630 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r2wr\" (UniqueName: \"kubernetes.io/projected/85b1d35a-e65a-47c7-afe9-6a6caf924106-kube-api-access-8r2wr\") pod \"openshift-apiserver-operator-796bbdcf4f-8htb6\" (UID: \"85b1d35a-e65a-47c7-afe9-6a6caf924106\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656650 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656667 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-etcd-serving-ca\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656686 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b1d35a-e65a-47c7-afe9-6a6caf924106-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8htb6\" (UID: \"85b1d35a-e65a-47c7-afe9-6a6caf924106\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656732 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dssg\" (UniqueName: \"kubernetes.io/projected/f00c55b0-60e2-4d35-a250-706ec8d9de8a-kube-api-access-4dssg\") pod \"cluster-samples-operator-665b6dd947-tr8hd\" (UID: \"f00c55b0-60e2-4d35-a250-706ec8d9de8a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656748 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d3be498a-3362-4c71-85ac-e53b4889eec5-etcd-client\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656765 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/130674aa-a02b-4d35-bf24-0c74de51fe4d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656784 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656801 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85b1d35a-e65a-47c7-afe9-6a6caf924106-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8htb6\" (UID: \"85b1d35a-e65a-47c7-afe9-6a6caf924106\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656824 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-client-ca\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656839 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3be498a-3362-4c71-85ac-e53b4889eec5-serving-cert\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656854 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/130674aa-a02b-4d35-bf24-0c74de51fe4d-serving-cert\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656887 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc45t\" (UniqueName: \"kubernetes.io/projected/34482250-9472-45fe-aab5-485c7f129bca-kube-api-access-nc45t\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656902 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-config\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656917 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh2vg\" (UniqueName: \"kubernetes.io/projected/7d0ff181-18dc-47c5-b173-d5197b923176-kube-api-access-rh2vg\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656933 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-config\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.656988 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fb3271e-4d33-4ef0-b269-ca69f698fb66-serving-cert\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657004 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wr2h\" (UniqueName: \"kubernetes.io/projected/a13a950e-dc37-4878-afdf-623a7c7d01b1-kube-api-access-9wr2h\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkgwz\" (UID: \"a13a950e-dc37-4878-afdf-623a7c7d01b1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657022 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657038 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-policies\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657056 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-audit\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657074 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hcmt\" (UniqueName: \"kubernetes.io/projected/130674aa-a02b-4d35-bf24-0c74de51fe4d-kube-api-access-5hcmt\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657089 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36865289-7421-41dd-8434-fb2b31a3b974-serving-cert\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657105 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657121 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/34482250-9472-45fe-aab5-485c7f129bca-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657137 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-image-import-ca\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657152 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm6c6\" (UniqueName: \"kubernetes.io/projected/36865289-7421-41dd-8434-fb2b31a3b974-kube-api-access-wm6c6\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657178 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d3be498a-3362-4c71-85ac-e53b4889eec5-encryption-config\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657202 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657218 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-config\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657234 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d0ff181-18dc-47c5-b173-d5197b923176-config\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657255 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-config\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657269 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-client-ca\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.657285 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d3be498a-3362-4c71-85ac-e53b4889eec5-node-pullsecrets\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.660173 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-dir\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.662783 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.662896 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.662998 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.663142 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.663201 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.663228 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.663427 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.663502 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-zjpgj"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.664047 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zjpgj" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.664381 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.664549 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.664747 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.665503 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.666901 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.671632 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.671882 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.672003 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.701453 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.703498 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.704210 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.705054 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.706738 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.707718 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.707892 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.709365 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.710063 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.716092 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.718399 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.718666 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.718758 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.718891 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.719099 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.719269 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.719422 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.719658 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.719830 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.720021 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.722148 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.723740 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.723788 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.724710 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-policies\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.728487 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.729360 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.729691 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.730799 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.733551 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.734723 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.736285 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.737026 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.741400 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.741908 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.742866 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.743839 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.745663 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.749046 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.749374 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.749855 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.750494 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.751215 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.752181 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.756660 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.762342 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b885v\" (UniqueName: \"kubernetes.io/projected/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-kube-api-access-b885v\") pod \"oauth-openshift-558db77b4-6r89v\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.776388 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777006 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7qbg\" (UniqueName: \"kubernetes.io/projected/27821f2f-df7e-42d2-b687-244c7b6e5a3f-kube-api-access-w7qbg\") pod \"openshift-config-operator-7777fb866f-qmtqd\" (UID: \"27821f2f-df7e-42d2-b687-244c7b6e5a3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777068 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fb3271e-4d33-4ef0-b269-ca69f698fb66-serving-cert\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777099 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wr2h\" (UniqueName: \"kubernetes.io/projected/a13a950e-dc37-4878-afdf-623a7c7d01b1-kube-api-access-9wr2h\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkgwz\" (UID: \"a13a950e-dc37-4878-afdf-623a7c7d01b1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777124 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-audit\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777148 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hcmt\" (UniqueName: \"kubernetes.io/projected/130674aa-a02b-4d35-bf24-0c74de51fe4d-kube-api-access-5hcmt\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777172 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/34482250-9472-45fe-aab5-485c7f129bca-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777200 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36865289-7421-41dd-8434-fb2b31a3b974-serving-cert\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777222 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm6c6\" (UniqueName: \"kubernetes.io/projected/36865289-7421-41dd-8434-fb2b31a3b974-kube-api-access-wm6c6\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777246 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-image-import-ca\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777271 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d3be498a-3362-4c71-85ac-e53b4889eec5-encryption-config\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777322 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-oauth-config\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777378 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-config\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777402 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-serving-cert\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777406 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777428 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d0ff181-18dc-47c5-b173-d5197b923176-config\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777470 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/27821f2f-df7e-42d2-b687-244c7b6e5a3f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qmtqd\" (UID: \"27821f2f-df7e-42d2-b687-244c7b6e5a3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777499 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3359cd7b-e283-4f9c-baf6-29e6e85e247e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777526 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d3be498a-3362-4c71-85ac-e53b4889eec5-node-pullsecrets\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777549 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-config\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777572 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-client-ca\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777598 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a13a950e-dc37-4878-afdf-623a7c7d01b1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkgwz\" (UID: \"a13a950e-dc37-4878-afdf-623a7c7d01b1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777624 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvkzc\" (UniqueName: \"kubernetes.io/projected/e74e438d-61ef-403a-b958-8704f195f0ce-kube-api-access-tvkzc\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777652 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130674aa-a02b-4d35-bf24-0c74de51fe4d-config\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777674 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpkt9\" (UniqueName: \"kubernetes.io/projected/3fb3271e-4d33-4ef0-b269-ca69f698fb66-kube-api-access-fpkt9\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777698 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-service-ca\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777722 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777746 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20355c26-2666-4915-87d3-b61747ed5692-serving-cert\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777769 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a13a950e-dc37-4878-afdf-623a7c7d01b1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkgwz\" (UID: \"a13a950e-dc37-4878-afdf-623a7c7d01b1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777793 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3359cd7b-e283-4f9c-baf6-29e6e85e247e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777818 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-oauth-serving-cert\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777854 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/130674aa-a02b-4d35-bf24-0c74de51fe4d-service-ca-bundle\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777878 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-images\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777902 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3359cd7b-e283-4f9c-baf6-29e6e85e247e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777926 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7d0ff181-18dc-47c5-b173-d5197b923176-auth-proxy-config\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777965 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7d0ff181-18dc-47c5-b173-d5197b923176-machine-approver-tls\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.777990 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f00c55b0-60e2-4d35-a250-706ec8d9de8a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tr8hd\" (UID: \"f00c55b0-60e2-4d35-a250-706ec8d9de8a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778014 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m559n\" (UniqueName: \"kubernetes.io/projected/20355c26-2666-4915-87d3-b61747ed5692-kube-api-access-m559n\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778036 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-console-config\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778061 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d3be498a-3362-4c71-85ac-e53b4889eec5-audit-dir\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778082 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x445g\" (UniqueName: \"kubernetes.io/projected/d3be498a-3362-4c71-85ac-e53b4889eec5-kube-api-access-x445g\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778131 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r2wr\" (UniqueName: \"kubernetes.io/projected/85b1d35a-e65a-47c7-afe9-6a6caf924106-kube-api-access-8r2wr\") pod \"openshift-apiserver-operator-796bbdcf4f-8htb6\" (UID: \"85b1d35a-e65a-47c7-afe9-6a6caf924106\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778149 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778154 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20355c26-2666-4915-87d3-b61747ed5692-trusted-ca\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778182 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-etcd-serving-ca\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778207 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b1d35a-e65a-47c7-afe9-6a6caf924106-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8htb6\" (UID: \"85b1d35a-e65a-47c7-afe9-6a6caf924106\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778232 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5afc2ea-9954-4617-bd56-5493fc53121f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qjxgp\" (UID: \"d5afc2ea-9954-4617-bd56-5493fc53121f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778255 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dssg\" (UniqueName: \"kubernetes.io/projected/f00c55b0-60e2-4d35-a250-706ec8d9de8a-kube-api-access-4dssg\") pod \"cluster-samples-operator-665b6dd947-tr8hd\" (UID: \"f00c55b0-60e2-4d35-a250-706ec8d9de8a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778279 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9k8g\" (UniqueName: \"kubernetes.io/projected/0faea847-63f9-488b-9f3a-080e36fd9025-kube-api-access-c9k8g\") pod \"downloads-7954f5f757-zjpgj\" (UID: \"0faea847-63f9-488b-9f3a-080e36fd9025\") " pod="openshift-console/downloads-7954f5f757-zjpgj" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778302 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5afc2ea-9954-4617-bd56-5493fc53121f-config\") pod \"kube-controller-manager-operator-78b949d7b-qjxgp\" (UID: \"d5afc2ea-9954-4617-bd56-5493fc53121f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778325 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27821f2f-df7e-42d2-b687-244c7b6e5a3f-serving-cert\") pod \"openshift-config-operator-7777fb866f-qmtqd\" (UID: \"27821f2f-df7e-42d2-b687-244c7b6e5a3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778347 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d5afc2ea-9954-4617-bd56-5493fc53121f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qjxgp\" (UID: \"d5afc2ea-9954-4617-bd56-5493fc53121f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778379 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d3be498a-3362-4c71-85ac-e53b4889eec5-etcd-client\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778407 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/130674aa-a02b-4d35-bf24-0c74de51fe4d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778459 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-client-ca\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778479 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85b1d35a-e65a-47c7-afe9-6a6caf924106-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8htb6\" (UID: \"85b1d35a-e65a-47c7-afe9-6a6caf924106\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778501 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9gfz\" (UniqueName: \"kubernetes.io/projected/1d593c77-f754-4464-bc94-5c265029d330-kube-api-access-m9gfz\") pod \"migrator-59844c95c7-9s6h2\" (UID: \"1d593c77-f754-4464-bc94-5c265029d330\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778526 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/130674aa-a02b-4d35-bf24-0c74de51fe4d-serving-cert\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778547 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20355c26-2666-4915-87d3-b61747ed5692-config\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778573 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3be498a-3362-4c71-85ac-e53b4889eec5-serving-cert\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778676 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778705 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc45t\" (UniqueName: \"kubernetes.io/projected/34482250-9472-45fe-aab5-485c7f129bca-kube-api-access-nc45t\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778729 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg4gz\" (UniqueName: \"kubernetes.io/projected/3359cd7b-e283-4f9c-baf6-29e6e85e247e-kube-api-access-bg4gz\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778752 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-config\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778773 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh2vg\" (UniqueName: \"kubernetes.io/projected/7d0ff181-18dc-47c5-b173-d5197b923176-kube-api-access-rh2vg\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778796 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-config\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.778830 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-trusted-ca-bundle\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.779156 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-client-ca\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.780669 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d0ff181-18dc-47c5-b173-d5197b923176-config\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.782651 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.784921 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-audit\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.785029 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/130674aa-a02b-4d35-bf24-0c74de51fe4d-service-ca-bundle\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.785367 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d3be498a-3362-4c71-85ac-e53b4889eec5-node-pullsecrets\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.786466 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7d0ff181-18dc-47c5-b173-d5197b923176-auth-proxy-config\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.787210 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-config\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.790154 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.791160 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.791438 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2w25g"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.791692 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9h4lw"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.791973 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.792319 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.792505 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.792702 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.792934 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.793167 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.793392 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.793628 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a13a950e-dc37-4878-afdf-623a7c7d01b1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkgwz\" (UID: \"a13a950e-dc37-4878-afdf-623a7c7d01b1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.793804 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.794375 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130674aa-a02b-4d35-bf24-0c74de51fe4d-config\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.799797 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-image-import-ca\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.799816 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rjhgs"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.801654 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d3be498a-3362-4c71-85ac-e53b4889eec5-audit-dir\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.801772 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7d0ff181-18dc-47c5-b173-d5197b923176-machine-approver-tls\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.801888 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-images\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.802082 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/130674aa-a02b-4d35-bf24-0c74de51fe4d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.802285 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-etcd-serving-ca\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.802752 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36865289-7421-41dd-8434-fb2b31a3b974-serving-cert\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.802757 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b1d35a-e65a-47c7-afe9-6a6caf924106-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8htb6\" (UID: \"85b1d35a-e65a-47c7-afe9-6a6caf924106\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.808363 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3be498a-3362-4c71-85ac-e53b4889eec5-serving-cert\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.809485 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85b1d35a-e65a-47c7-afe9-6a6caf924106-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8htb6\" (UID: \"85b1d35a-e65a-47c7-afe9-6a6caf924106\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.809519 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-config\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.809744 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r89v"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.809779 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.810587 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.811431 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3be498a-3362-4c71-85ac-e53b4889eec5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.812023 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.812309 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-client-ca\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.814511 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wr2h\" (UniqueName: \"kubernetes.io/projected/a13a950e-dc37-4878-afdf-623a7c7d01b1-kube-api-access-9wr2h\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkgwz\" (UID: \"a13a950e-dc37-4878-afdf-623a7c7d01b1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.817607 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.818221 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d3be498a-3362-4c71-85ac-e53b4889eec5-etcd-client\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.821854 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-config\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.822020 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.822158 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/130674aa-a02b-4d35-bf24-0c74de51fe4d-serving-cert\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.822753 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bnwh2"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.823270 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.823272 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.824054 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dwhtv"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.825421 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.828278 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f00c55b0-60e2-4d35-a250-706ec8d9de8a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tr8hd\" (UID: \"f00c55b0-60e2-4d35-a250-706ec8d9de8a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.829396 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hcmt\" (UniqueName: \"kubernetes.io/projected/130674aa-a02b-4d35-bf24-0c74de51fe4d-kube-api-access-5hcmt\") pod \"authentication-operator-69f744f599-8nz76\" (UID: \"130674aa-a02b-4d35-bf24-0c74de51fe4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.830209 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpkt9\" (UniqueName: \"kubernetes.io/projected/3fb3271e-4d33-4ef0-b269-ca69f698fb66-kube-api-access-fpkt9\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.831221 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm6c6\" (UniqueName: \"kubernetes.io/projected/36865289-7421-41dd-8434-fb2b31a3b974-kube-api-access-wm6c6\") pod \"controller-manager-879f6c89f-jfpmq\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.833117 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fb3271e-4d33-4ef0-b269-ca69f698fb66-serving-cert\") pod \"route-controller-manager-6576b87f9c-tgtkh\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.833900 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lj984"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.834645 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.836160 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a13a950e-dc37-4878-afdf-623a7c7d01b1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkgwz\" (UID: \"a13a950e-dc37-4878-afdf-623a7c7d01b1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.837899 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.838113 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.838348 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.839895 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jbnv7"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.841024 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.841154 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d3be498a-3362-4c71-85ac-e53b4889eec5-encryption-config\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.841775 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.843091 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.844317 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vspww"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.845084 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.846377 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-527r8"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.846768 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.847758 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.849773 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.849898 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.850531 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.851748 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.853101 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.853389 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.854574 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.855102 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.855126 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.855647 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.855869 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.856996 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.862638 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.865314 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2hhv9"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.866337 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2hhv9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.875897 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.879642 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.880184 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m559n\" (UniqueName: \"kubernetes.io/projected/20355c26-2666-4915-87d3-b61747ed5692-kube-api-access-m559n\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.880376 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-etcd-client\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.880554 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jqvl\" (UniqueName: \"kubernetes.io/projected/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-kube-api-access-2jqvl\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.880860 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f17ca838-95f0-4516-acf0-4d896c5be737-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-thglp\" (UID: \"f17ca838-95f0-4516-acf0-4d896c5be737\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.880914 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ada6c8-9f70-4509-ad61-d31fb216090b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dwhtv\" (UID: \"00ada6c8-9f70-4509-ad61-d31fb216090b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.880932 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7ad4e51a-8091-4085-8c59-2a15dcc9a0ec-proxy-tls\") pod \"machine-config-controller-84d6567774-2jsgc\" (UID: \"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.881037 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.881478 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.882112 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.882335 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9k8g\" (UniqueName: \"kubernetes.io/projected/0faea847-63f9-488b-9f3a-080e36fd9025-kube-api-access-c9k8g\") pod \"downloads-7954f5f757-zjpgj\" (UID: \"0faea847-63f9-488b-9f3a-080e36fd9025\") " pod="openshift-console/downloads-7954f5f757-zjpgj" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.882441 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5afc2ea-9954-4617-bd56-5493fc53121f-config\") pod \"kube-controller-manager-operator-78b949d7b-qjxgp\" (UID: \"d5afc2ea-9954-4617-bd56-5493fc53121f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.882528 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85tdz\" (UniqueName: \"kubernetes.io/projected/82f68827-538c-421d-ab03-2e0834d22429-kube-api-access-85tdz\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.882630 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d5afc2ea-9954-4617-bd56-5493fc53121f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qjxgp\" (UID: \"d5afc2ea-9954-4617-bd56-5493fc53121f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.882798 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9gfz\" (UniqueName: \"kubernetes.io/projected/1d593c77-f754-4464-bc94-5c265029d330-kube-api-access-m9gfz\") pod \"migrator-59844c95c7-9s6h2\" (UID: \"1d593c77-f754-4464-bc94-5c265029d330\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.882904 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20355c26-2666-4915-87d3-b61747ed5692-config\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883052 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg4gz\" (UniqueName: \"kubernetes.io/projected/3359cd7b-e283-4f9c-baf6-29e6e85e247e-kube-api-access-bg4gz\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883189 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjm5b\" (UniqueName: \"kubernetes.io/projected/a930872e-2679-46c9-ac6f-22bffe1b8d66-kube-api-access-hjm5b\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883343 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5afc2ea-9954-4617-bd56-5493fc53121f-config\") pod \"kube-controller-manager-operator-78b949d7b-qjxgp\" (UID: \"d5afc2ea-9954-4617-bd56-5493fc53121f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883310 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-encryption-config\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883527 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmv2g\" (UniqueName: \"kubernetes.io/projected/61e54f93-d45e-47f0-b610-acd60abbf88c-kube-api-access-fmv2g\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883583 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a930872e-2679-46c9-ac6f-22bffe1b8d66-secret-volume\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883644 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/96ca6b03-36ef-47ec-9145-5b45994312cf-tmpfs\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883688 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-oauth-config\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883740 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-serving-cert\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883824 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883860 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/27821f2f-df7e-42d2-b687-244c7b6e5a3f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qmtqd\" (UID: \"27821f2f-df7e-42d2-b687-244c7b6e5a3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.883964 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-serving-cert\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884019 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvkzc\" (UniqueName: \"kubernetes.io/projected/e74e438d-61ef-403a-b958-8704f195f0ce-kube-api-access-tvkzc\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884055 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-srv-cert\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884081 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-audit-dir\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884129 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20355c26-2666-4915-87d3-b61747ed5692-serving-cert\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884169 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-profile-collector-cert\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884237 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3359cd7b-e283-4f9c-baf6-29e6e85e247e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884269 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-audit-policies\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884321 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-console-config\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884351 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5cd27419-02e2-4228-b9f6-e6727c69722b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-v7mzq\" (UID: \"5cd27419-02e2-4228-b9f6-e6727c69722b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884071 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20355c26-2666-4915-87d3-b61747ed5692-config\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.884483 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20355c26-2666-4915-87d3-b61747ed5692-trusted-ca\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.887230 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zjpgj"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.888117 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/27821f2f-df7e-42d2-b687-244c7b6e5a3f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qmtqd\" (UID: \"27821f2f-df7e-42d2-b687-244c7b6e5a3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.890169 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5afc2ea-9954-4617-bd56-5493fc53121f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qjxgp\" (UID: \"d5afc2ea-9954-4617-bd56-5493fc53121f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.890233 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f17ca838-95f0-4516-acf0-4d896c5be737-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-thglp\" (UID: \"f17ca838-95f0-4516-acf0-4d896c5be737\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.891401 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k64kb"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.892103 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-images\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.892251 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f68827-538c-421d-ab03-2e0834d22429-config\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.892281 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27821f2f-df7e-42d2-b687-244c7b6e5a3f-serving-cert\") pod \"openshift-config-operator-7777fb866f-qmtqd\" (UID: \"27821f2f-df7e-42d2-b687-244c7b6e5a3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.892301 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-key\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.892320 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.892379 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-webhook-cert\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.892426 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-trusted-ca-bundle\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.892448 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-proxy-tls\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.892492 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7qbg\" (UniqueName: \"kubernetes.io/projected/27821f2f-df7e-42d2-b687-244c7b6e5a3f-kube-api-access-w7qbg\") pod \"openshift-config-operator-7777fb866f-qmtqd\" (UID: \"27821f2f-df7e-42d2-b687-244c7b6e5a3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893045 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh2gq\" (UniqueName: \"kubernetes.io/projected/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-kube-api-access-jh2gq\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893163 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dfcj\" (UniqueName: \"kubernetes.io/projected/1626b474-7e91-4cde-bab7-a4db753478dd-kube-api-access-9dfcj\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893244 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnfj6\" (UniqueName: \"kubernetes.io/projected/7ad4e51a-8091-4085-8c59-2a15dcc9a0ec-kube-api-access-fnfj6\") pod \"machine-config-controller-84d6567774-2jsgc\" (UID: \"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893287 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ad4e51a-8091-4085-8c59-2a15dcc9a0ec-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2jsgc\" (UID: \"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893310 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2g69\" (UniqueName: \"kubernetes.io/projected/00ada6c8-9f70-4509-ad61-d31fb216090b-kube-api-access-c2g69\") pod \"multus-admission-controller-857f4d67dd-dwhtv\" (UID: \"00ada6c8-9f70-4509-ad61-d31fb216090b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893335 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x76q8\" (UniqueName: \"kubernetes.io/projected/5cd27419-02e2-4228-b9f6-e6727c69722b-kube-api-access-x76q8\") pod \"package-server-manager-789f6589d5-v7mzq\" (UID: \"5cd27419-02e2-4228-b9f6-e6727c69722b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893360 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3359cd7b-e283-4f9c-baf6-29e6e85e247e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893381 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893403 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82f68827-538c-421d-ab03-2e0834d22429-serving-cert\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893430 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q85b\" (UniqueName: \"kubernetes.io/projected/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-kube-api-access-7q85b\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893448 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-apiservice-cert\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893468 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a930872e-2679-46c9-ac6f-22bffe1b8d66-config-volume\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893504 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-service-ca\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893525 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f17ca838-95f0-4516-acf0-4d896c5be737-config\") pod \"kube-apiserver-operator-766d6c64bb-thglp\" (UID: \"f17ca838-95f0-4516-acf0-4d896c5be737\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893544 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z6lf\" (UniqueName: \"kubernetes.io/projected/96ca6b03-36ef-47ec-9145-5b45994312cf-kube-api-access-2z6lf\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893569 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3359cd7b-e283-4f9c-baf6-29e6e85e247e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893589 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-oauth-serving-cert\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.893630 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-cabundle\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.895226 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-console-config\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.895467 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-service-ca\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.895836 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20355c26-2666-4915-87d3-b61747ed5692-trusted-ca\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.896569 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-trusted-ca-bundle\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.897177 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20355c26-2666-4915-87d3-b61747ed5692-serving-cert\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.897470 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3359cd7b-e283-4f9c-baf6-29e6e85e247e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.898104 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-oauth-serving-cert\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.899176 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.899855 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-serving-cert\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.900968 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-oauth-config\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.901234 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5afc2ea-9954-4617-bd56-5493fc53121f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qjxgp\" (UID: \"d5afc2ea-9954-4617-bd56-5493fc53121f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.902225 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jfpmq"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.902259 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.902271 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.902283 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-l8sf2"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.903290 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.903538 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27821f2f-df7e-42d2-b687-244c7b6e5a3f-serving-cert\") pod \"openshift-config-operator-7777fb866f-qmtqd\" (UID: \"27821f2f-df7e-42d2-b687-244c7b6e5a3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.903204 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lkzd9"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.904556 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8nz76"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.905826 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.906006 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.908435 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.909330 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3359cd7b-e283-4f9c-baf6-29e6e85e247e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.909525 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vspww"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.910601 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.911851 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.912736 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bnwh2"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.913871 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-x4gh2"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.915586 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.917437 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2w25g"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.918195 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.918221 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.919752 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.922632 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.923874 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.925408 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.927807 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.928702 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-f29nb"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.929590 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.929759 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dwhtv"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.931576 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9h4lw"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.932993 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.933646 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2hhv9"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.936779 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.937418 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.938067 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.938845 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-527r8"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.940245 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.943124 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-l8sf2"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.944543 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.946042 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.947013 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.959744 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.962900 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.964138 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lkzd9"] Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.982801 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998187 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f17ca838-95f0-4516-acf0-4d896c5be737-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-thglp\" (UID: \"f17ca838-95f0-4516-acf0-4d896c5be737\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998223 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-images\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998250 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f68827-538c-421d-ab03-2e0834d22429-config\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998268 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-key\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998286 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998320 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-webhook-cert\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998358 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-proxy-tls\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998390 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh2gq\" (UniqueName: \"kubernetes.io/projected/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-kube-api-access-jh2gq\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998409 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dfcj\" (UniqueName: \"kubernetes.io/projected/1626b474-7e91-4cde-bab7-a4db753478dd-kube-api-access-9dfcj\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998436 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnfj6\" (UniqueName: \"kubernetes.io/projected/7ad4e51a-8091-4085-8c59-2a15dcc9a0ec-kube-api-access-fnfj6\") pod \"machine-config-controller-84d6567774-2jsgc\" (UID: \"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998457 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2g69\" (UniqueName: \"kubernetes.io/projected/00ada6c8-9f70-4509-ad61-d31fb216090b-kube-api-access-c2g69\") pod \"multus-admission-controller-857f4d67dd-dwhtv\" (UID: \"00ada6c8-9f70-4509-ad61-d31fb216090b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998476 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ad4e51a-8091-4085-8c59-2a15dcc9a0ec-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2jsgc\" (UID: \"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998495 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x76q8\" (UniqueName: \"kubernetes.io/projected/5cd27419-02e2-4228-b9f6-e6727c69722b-kube-api-access-x76q8\") pod \"package-server-manager-789f6589d5-v7mzq\" (UID: \"5cd27419-02e2-4228-b9f6-e6727c69722b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998527 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998550 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82f68827-538c-421d-ab03-2e0834d22429-serving-cert\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998570 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q85b\" (UniqueName: \"kubernetes.io/projected/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-kube-api-access-7q85b\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998591 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-apiservice-cert\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998612 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a930872e-2679-46c9-ac6f-22bffe1b8d66-config-volume\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998629 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f17ca838-95f0-4516-acf0-4d896c5be737-config\") pod \"kube-apiserver-operator-766d6c64bb-thglp\" (UID: \"f17ca838-95f0-4516-acf0-4d896c5be737\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998643 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z6lf\" (UniqueName: \"kubernetes.io/projected/96ca6b03-36ef-47ec-9145-5b45994312cf-kube-api-access-2z6lf\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998664 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-cabundle\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998685 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-etcd-client\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998702 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jqvl\" (UniqueName: \"kubernetes.io/projected/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-kube-api-access-2jqvl\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998727 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f17ca838-95f0-4516-acf0-4d896c5be737-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-thglp\" (UID: \"f17ca838-95f0-4516-acf0-4d896c5be737\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998748 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ada6c8-9f70-4509-ad61-d31fb216090b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dwhtv\" (UID: \"00ada6c8-9f70-4509-ad61-d31fb216090b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998766 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7ad4e51a-8091-4085-8c59-2a15dcc9a0ec-proxy-tls\") pod \"machine-config-controller-84d6567774-2jsgc\" (UID: \"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998790 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998809 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998824 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85tdz\" (UniqueName: \"kubernetes.io/projected/82f68827-538c-421d-ab03-2e0834d22429-kube-api-access-85tdz\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998861 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjm5b\" (UniqueName: \"kubernetes.io/projected/a930872e-2679-46c9-ac6f-22bffe1b8d66-kube-api-access-hjm5b\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998876 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-encryption-config\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998892 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmv2g\" (UniqueName: \"kubernetes.io/projected/61e54f93-d45e-47f0-b610-acd60abbf88c-kube-api-access-fmv2g\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998923 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a930872e-2679-46c9-ac6f-22bffe1b8d66-secret-volume\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.998982 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/96ca6b03-36ef-47ec-9145-5b45994312cf-tmpfs\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.999017 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.999032 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-serving-cert\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:52 crc kubenswrapper[4735]: I0215 20:18:52.999056 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-srv-cert\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.004755 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-audit-dir\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.004795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-profile-collector-cert\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.004830 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-audit-policies\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.004858 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5cd27419-02e2-4228-b9f6-e6727c69722b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-v7mzq\" (UID: \"5cd27419-02e2-4228-b9f6-e6727c69722b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.004025 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-images\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.001872 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.005018 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-audit-dir\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.003048 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.003367 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.004008 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/96ca6b03-36ef-47ec-9145-5b45994312cf-tmpfs\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.005987 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f17ca838-95f0-4516-acf0-4d896c5be737-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-thglp\" (UID: \"f17ca838-95f0-4516-acf0-4d896c5be737\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.007045 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ad4e51a-8091-4085-8c59-2a15dcc9a0ec-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2jsgc\" (UID: \"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.009744 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-proxy-tls\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.011742 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f17ca838-95f0-4516-acf0-4d896c5be737-config\") pod \"kube-apiserver-operator-766d6c64bb-thglp\" (UID: \"f17ca838-95f0-4516-acf0-4d896c5be737\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.019099 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.042263 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.056926 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-etcd-client\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.057543 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.062971 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.070598 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-serving-cert\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.078456 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.092111 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-encryption-config\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.103354 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.110010 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.110877 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.118872 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.137668 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r89v"] Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.148903 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.157697 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.166203 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-audit-policies\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.179001 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.197715 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.221319 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.228249 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8nz76"] Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.238277 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.259638 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.278857 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.290313 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz"] Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.299386 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.319222 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.339472 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.348106 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jfpmq"] Feb 15 20:18:53 crc kubenswrapper[4735]: W0215 20:18:53.357474 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36865289_7421_41dd_8434_fb2b31a3b974.slice/crio-84e08b76e7809727f99d9c58c0aad2562c1dc8394a82860ff101b489d6328037 WatchSource:0}: Error finding container 84e08b76e7809727f99d9c58c0aad2562c1dc8394a82860ff101b489d6328037: Status 404 returned error can't find the container with id 84e08b76e7809727f99d9c58c0aad2562c1dc8394a82860ff101b489d6328037 Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.358765 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.378145 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.392605 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh"] Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.398397 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.418461 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.478319 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r2wr\" (UniqueName: \"kubernetes.io/projected/85b1d35a-e65a-47c7-afe9-6a6caf924106-kube-api-access-8r2wr\") pod \"openshift-apiserver-operator-796bbdcf4f-8htb6\" (UID: \"85b1d35a-e65a-47c7-afe9-6a6caf924106\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.494427 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dssg\" (UniqueName: \"kubernetes.io/projected/f00c55b0-60e2-4d35-a250-706ec8d9de8a-kube-api-access-4dssg\") pod \"cluster-samples-operator-665b6dd947-tr8hd\" (UID: \"f00c55b0-60e2-4d35-a250-706ec8d9de8a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.545807 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.547311 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.558574 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.578922 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.598954 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.619596 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.644000 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.658879 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.659625 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh2vg\" (UniqueName: \"kubernetes.io/projected/7d0ff181-18dc-47c5-b173-d5197b923176-kube-api-access-rh2vg\") pod \"machine-approver-56656f9798-fvv22\" (UID: \"7d0ff181-18dc-47c5-b173-d5197b923176\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.696226 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.698201 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.698216 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.714190 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7ad4e51a-8091-4085-8c59-2a15dcc9a0ec-proxy-tls\") pod \"machine-config-controller-84d6567774-2jsgc\" (UID: \"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:53 crc kubenswrapper[4735]: W0215 20:18:53.718360 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d0ff181_18dc_47c5_b173_d5197b923176.slice/crio-9d902060a6b2281eb9d6019efc5ed20b6c828854865ada853f3d59f1957cf1f4 WatchSource:0}: Error finding container 9d902060a6b2281eb9d6019efc5ed20b6c828854865ada853f3d59f1957cf1f4: Status 404 returned error can't find the container with id 9d902060a6b2281eb9d6019efc5ed20b6c828854865ada853f3d59f1957cf1f4 Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.718561 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.738360 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.758550 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 15 20:18:53 crc kubenswrapper[4735]: E0215 20:18:53.781659 4735 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:53 crc kubenswrapper[4735]: E0215 20:18:53.781753 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-config podName:34482250-9472-45fe-aab5-485c7f129bca nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.281727657 +0000 UTC m=+142.147743280 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-config") pod "machine-api-operator-5694c8668f-rjhgs" (UID: "34482250-9472-45fe-aab5-485c7f129bca") : failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:53 crc kubenswrapper[4735]: E0215 20:18:53.781776 4735 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:53 crc kubenswrapper[4735]: E0215 20:18:53.781802 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34482250-9472-45fe-aab5-485c7f129bca-machine-api-operator-tls podName:34482250-9472-45fe-aab5-485c7f129bca nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.281794979 +0000 UTC m=+142.147810602 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/34482250-9472-45fe-aab5-485c7f129bca-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-rjhgs" (UID: "34482250-9472-45fe-aab5-485c7f129bca") : failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.784241 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6"] Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.785476 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.798395 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.804972 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ada6c8-9f70-4509-ad61-d31fb216090b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dwhtv\" (UID: \"00ada6c8-9f70-4509-ad61-d31fb216090b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.818232 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.835368 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" event={"ID":"36865289-7421-41dd-8434-fb2b31a3b974","Type":"ContainerStarted","Data":"1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.835431 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" event={"ID":"36865289-7421-41dd-8434-fb2b31a3b974","Type":"ContainerStarted","Data":"84e08b76e7809727f99d9c58c0aad2562c1dc8394a82860ff101b489d6328037"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.836472 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.836534 4735 request.go:700] Waited for 1.001570613s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-metrics-certs-default&limit=500&resourceVersion=0 Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.842707 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.846749 4735 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-jfpmq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.846785 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" podUID="36865289-7421-41dd-8434-fb2b31a3b974" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.857909 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.878258 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.880423 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" event={"ID":"9947c023-d2cc-4e2d-989c-fec1d09e4ca8","Type":"ContainerStarted","Data":"56d37d29714a841c4bc005dbdd55124981688ee7a317302eb6bd82575d03c4a6"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.880472 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" event={"ID":"9947c023-d2cc-4e2d-989c-fec1d09e4ca8","Type":"ContainerStarted","Data":"192225447fdf282be618c96ec6af9a43a5e04e00941b397d9cbaef962138a803"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.881663 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.883166 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" event={"ID":"85b1d35a-e65a-47c7-afe9-6a6caf924106","Type":"ContainerStarted","Data":"e330758be5fe463a39ce533ce1fc28f1d575acc8c3ead282f02984efc37ca943"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.883961 4735 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6r89v container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.5:6443/healthz\": dial tcp 10.217.0.5:6443: connect: connection refused" start-of-body= Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.883997 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" podUID="9947c023-d2cc-4e2d-989c-fec1d09e4ca8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.5:6443/healthz\": dial tcp 10.217.0.5:6443: connect: connection refused" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.897915 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.899828 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" event={"ID":"3fb3271e-4d33-4ef0-b269-ca69f698fb66","Type":"ContainerStarted","Data":"2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.899864 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" event={"ID":"3fb3271e-4d33-4ef0-b269-ca69f698fb66","Type":"ContainerStarted","Data":"07f83cd951594661fd420167099f0926e8c47b3017b98a3b3498a4bff3398538"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.900910 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.914793 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" event={"ID":"7d0ff181-18dc-47c5-b173-d5197b923176","Type":"ContainerStarted","Data":"9d902060a6b2281eb9d6019efc5ed20b6c828854865ada853f3d59f1957cf1f4"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.914919 4735 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-tgtkh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.914982 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" podUID="3fb3271e-4d33-4ef0-b269-ca69f698fb66" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.919692 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.919772 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd"] Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.922339 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" event={"ID":"a13a950e-dc37-4878-afdf-623a7c7d01b1","Type":"ContainerStarted","Data":"87120cdee8cb8527926295f89523e531c8b082770e5d56e9ee45c021aedc3a21"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.922381 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" event={"ID":"a13a950e-dc37-4878-afdf-623a7c7d01b1","Type":"ContainerStarted","Data":"c5157a2a748f243501edab35c5e82bb6cc8aca8d0fb8e72b5e0f3b08c81739d6"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.926686 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" event={"ID":"130674aa-a02b-4d35-bf24-0c74de51fe4d","Type":"ContainerStarted","Data":"439256c656dc9e000a2f49a7cda912ae03cffbfd27fdebef9f834ff2c3a8b13e"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.926715 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" event={"ID":"130674aa-a02b-4d35-bf24-0c74de51fe4d","Type":"ContainerStarted","Data":"740ae7ea83ca77e24b546b4d22c2f0f0afa401bebe40c9339d96ae9ded1204ad"} Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.937675 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.958241 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.979333 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 15 20:18:53 crc kubenswrapper[4735]: I0215 20:18:53.998601 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 15 20:18:53 crc kubenswrapper[4735]: E0215 20:18:53.999546 4735 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:53 crc kubenswrapper[4735]: E0215 20:18:53.999630 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a930872e-2679-46c9-ac6f-22bffe1b8d66-config-volume podName:a930872e-2679-46c9-ac6f-22bffe1b8d66 nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.499610085 +0000 UTC m=+142.365625708 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/a930872e-2679-46c9-ac6f-22bffe1b8d66-config-volume") pod "collect-profiles-29519775-4dm4h" (UID: "a930872e-2679-46c9-ac6f-22bffe1b8d66") : failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.001374 4735 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.001473 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-cabundle podName:61e54f93-d45e-47f0-b610-acd60abbf88c nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.501452118 +0000 UTC m=+142.367467741 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-cabundle") pod "service-ca-9c57cc56f-vspww" (UID: "61e54f93-d45e-47f0-b610-acd60abbf88c") : failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.003440 4735 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.003531 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a930872e-2679-46c9-ac6f-22bffe1b8d66-secret-volume podName:a930872e-2679-46c9-ac6f-22bffe1b8d66 nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.503519988 +0000 UTC m=+142.369535611 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/a930872e-2679-46c9-ac6f-22bffe1b8d66-secret-volume") pod "collect-profiles-29519775-4dm4h" (UID: "a930872e-2679-46c9-ac6f-22bffe1b8d66") : failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.003709 4735 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.003763 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-operator-metrics podName:1626b474-7e91-4cde-bab7-a4db753478dd nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.503735224 +0000 UTC m=+142.369750847 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-operator-metrics") pod "marketplace-operator-79b997595-527r8" (UID: "1626b474-7e91-4cde-bab7-a4db753478dd") : failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005007 4735 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005034 4735 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005087 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-key podName:61e54f93-d45e-47f0-b610-acd60abbf88c nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.505073102 +0000 UTC m=+142.371088915 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-key") pod "service-ca-9c57cc56f-vspww" (UID: "61e54f93-d45e-47f0-b610-acd60abbf88c") : failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005108 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-webhook-cert podName:96ca6b03-36ef-47ec-9145-5b45994312cf nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.505095303 +0000 UTC m=+142.371110926 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-webhook-cert") pod "packageserver-d55dfcdfc-hm5c9" (UID: "96ca6b03-36ef-47ec-9145-5b45994312cf") : failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005123 4735 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005148 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82f68827-538c-421d-ab03-2e0834d22429-config podName:82f68827-538c-421d-ab03-2e0834d22429 nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.505141344 +0000 UTC m=+142.371156967 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/82f68827-538c-421d-ab03-2e0834d22429-config") pod "service-ca-operator-777779d784-dg5fw" (UID: "82f68827-538c-421d-ab03-2e0834d22429") : failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005152 4735 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005178 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-trusted-ca podName:1626b474-7e91-4cde-bab7-a4db753478dd nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.505172225 +0000 UTC m=+142.371187848 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-trusted-ca") pod "marketplace-operator-79b997595-527r8" (UID: "1626b474-7e91-4cde-bab7-a4db753478dd") : failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005374 4735 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005410 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cd27419-02e2-4228-b9f6-e6727c69722b-package-server-manager-serving-cert podName:5cd27419-02e2-4228-b9f6-e6727c69722b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.505399831 +0000 UTC m=+142.371415654 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/5cd27419-02e2-4228-b9f6-e6727c69722b-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-v7mzq" (UID: "5cd27419-02e2-4228-b9f6-e6727c69722b") : failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005728 4735 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005762 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-apiservice-cert podName:96ca6b03-36ef-47ec-9145-5b45994312cf nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.505753322 +0000 UTC m=+142.371769145 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-apiservice-cert") pod "packageserver-d55dfcdfc-hm5c9" (UID: "96ca6b03-36ef-47ec-9145-5b45994312cf") : failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005785 4735 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005814 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82f68827-538c-421d-ab03-2e0834d22429-serving-cert podName:82f68827-538c-421d-ab03-2e0834d22429 nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.505806043 +0000 UTC m=+142.371821886 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82f68827-538c-421d-ab03-2e0834d22429-serving-cert") pod "service-ca-operator-777779d784-dg5fw" (UID: "82f68827-538c-421d-ab03-2e0834d22429") : failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005843 4735 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.005869 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-profile-collector-cert podName:9ec3f4d1-68c8-4460-8802-bc4ac3c1014f nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.505863195 +0000 UTC m=+142.371878818 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-profile-collector-cert") pod "catalog-operator-68c6474976-nhpdn" (UID: "9ec3f4d1-68c8-4460-8802-bc4ac3c1014f") : failed to sync secret cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.007474 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-srv-cert\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.021420 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.039280 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.058714 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.091287 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.097578 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.119602 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.141587 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.159596 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.194307 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.199558 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.219841 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.238445 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.258104 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.278997 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.298523 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.317980 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.339139 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.344094 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/34482250-9472-45fe-aab5-485c7f129bca-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.344141 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-config\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.359264 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.378571 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.398300 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.418042 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.438184 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.453957 4735 projected.go:288] Couldn't get configMap openshift-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.454014 4735 projected.go:194] Error preparing data for projected volume kube-api-access-x445g for pod openshift-apiserver/apiserver-76f77b778f-jbnv7: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.454074 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d3be498a-3362-4c71-85ac-e53b4889eec5-kube-api-access-x445g podName:d3be498a-3362-4c71-85ac-e53b4889eec5 nodeName:}" failed. No retries permitted until 2026-02-15 20:18:54.954054378 +0000 UTC m=+142.820070001 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-x445g" (UniqueName: "kubernetes.io/projected/d3be498a-3362-4c71-85ac-e53b4889eec5-kube-api-access-x445g") pod "apiserver-76f77b778f-jbnv7" (UID: "d3be498a-3362-4c71-85ac-e53b4889eec5") : failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.457828 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.477923 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.498146 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.510013 4735 projected.go:288] Couldn't get configMap openshift-machine-api/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.510044 4735 projected.go:194] Error preparing data for projected volume kube-api-access-nc45t for pod openshift-machine-api/machine-api-operator-5694c8668f-rjhgs: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: E0215 20:18:54.510112 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/34482250-9472-45fe-aab5-485c7f129bca-kube-api-access-nc45t podName:34482250-9472-45fe-aab5-485c7f129bca nodeName:}" failed. No retries permitted until 2026-02-15 20:18:55.01008713 +0000 UTC m=+142.876102763 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-nc45t" (UniqueName: "kubernetes.io/projected/34482250-9472-45fe-aab5-485c7f129bca-kube-api-access-nc45t") pod "machine-api-operator-5694c8668f-rjhgs" (UID: "34482250-9472-45fe-aab5-485c7f129bca") : failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.517922 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.538653 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547400 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547479 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-profile-collector-cert\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547517 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5cd27419-02e2-4228-b9f6-e6727c69722b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-v7mzq\" (UID: \"5cd27419-02e2-4228-b9f6-e6727c69722b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547539 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f68827-538c-421d-ab03-2e0834d22429-config\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547558 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547574 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-key\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547602 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-webhook-cert\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547678 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82f68827-538c-421d-ab03-2e0834d22429-serving-cert\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547714 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-apiservice-cert\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547736 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a930872e-2679-46c9-ac6f-22bffe1b8d66-config-volume\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547763 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-cabundle\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.547850 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a930872e-2679-46c9-ac6f-22bffe1b8d66-secret-volume\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.549590 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f68827-538c-421d-ab03-2e0834d22429-config\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.552787 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-webhook-cert\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.553116 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5cd27419-02e2-4228-b9f6-e6727c69722b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-v7mzq\" (UID: \"5cd27419-02e2-4228-b9f6-e6727c69722b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.553184 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.553854 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a930872e-2679-46c9-ac6f-22bffe1b8d66-secret-volume\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.553962 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a930872e-2679-46c9-ac6f-22bffe1b8d66-config-volume\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.554137 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-cabundle\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.554205 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-profile-collector-cert\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.555427 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.558283 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.558481 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82f68827-538c-421d-ab03-2e0834d22429-serving-cert\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.558561 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96ca6b03-36ef-47ec-9145-5b45994312cf-apiservice-cert\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.561000 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/61e54f93-d45e-47f0-b610-acd60abbf88c-signing-key\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.598754 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.621372 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.640682 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.658629 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.703013 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m559n\" (UniqueName: \"kubernetes.io/projected/20355c26-2666-4915-87d3-b61747ed5692-kube-api-access-m559n\") pod \"console-operator-58897d9998-x4gh2\" (UID: \"20355c26-2666-4915-87d3-b61747ed5692\") " pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.723633 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9k8g\" (UniqueName: \"kubernetes.io/projected/0faea847-63f9-488b-9f3a-080e36fd9025-kube-api-access-c9k8g\") pod \"downloads-7954f5f757-zjpgj\" (UID: \"0faea847-63f9-488b-9f3a-080e36fd9025\") " pod="openshift-console/downloads-7954f5f757-zjpgj" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.741798 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d5afc2ea-9954-4617-bd56-5493fc53121f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qjxgp\" (UID: \"d5afc2ea-9954-4617-bd56-5493fc53121f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.762563 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9gfz\" (UniqueName: \"kubernetes.io/projected/1d593c77-f754-4464-bc94-5c265029d330-kube-api-access-m9gfz\") pod \"migrator-59844c95c7-9s6h2\" (UID: \"1d593c77-f754-4464-bc94-5c265029d330\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.777809 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvkzc\" (UniqueName: \"kubernetes.io/projected/e74e438d-61ef-403a-b958-8704f195f0ce-kube-api-access-tvkzc\") pod \"console-f9d7485db-k64kb\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.789630 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.810762 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg4gz\" (UniqueName: \"kubernetes.io/projected/3359cd7b-e283-4f9c-baf6-29e6e85e247e-kube-api-access-bg4gz\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.834612 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7qbg\" (UniqueName: \"kubernetes.io/projected/27821f2f-df7e-42d2-b687-244c7b6e5a3f-kube-api-access-w7qbg\") pod \"openshift-config-operator-7777fb866f-qmtqd\" (UID: \"27821f2f-df7e-42d2-b687-244c7b6e5a3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.838527 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.853495 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3359cd7b-e283-4f9c-baf6-29e6e85e247e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-gd7bq\" (UID: \"3359cd7b-e283-4f9c-baf6-29e6e85e247e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.856261 4735 request.go:700] Waited for 1.952674462s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0 Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.857626 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.897932 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.899861 4735 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.918029 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.933737 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" event={"ID":"85b1d35a-e65a-47c7-afe9-6a6caf924106","Type":"ContainerStarted","Data":"157e935f1b18eccf86035218b6ed8db073b62a54ce8b43f352412af5e7e405a0"} Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.935355 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" event={"ID":"f00c55b0-60e2-4d35-a250-706ec8d9de8a","Type":"ContainerStarted","Data":"0b6617cca81ec1e5ed424ee4ece4882dd5bf86b630d18d7ec661e4a4abda772b"} Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.935409 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" event={"ID":"f00c55b0-60e2-4d35-a250-706ec8d9de8a","Type":"ContainerStarted","Data":"ba6bbc0d93b7f6aac6067b15d3ba996c51333e41876f9aea91110797634f9aa5"} Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.935422 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" event={"ID":"f00c55b0-60e2-4d35-a250-706ec8d9de8a","Type":"ContainerStarted","Data":"5850311df237f5e3db9451e5e261aa00bee234f2c004678b361605f7492778d4"} Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.937591 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.937654 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.941671 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" event={"ID":"7d0ff181-18dc-47c5-b173-d5197b923176","Type":"ContainerStarted","Data":"9787b162f0137af01461dae1a2222e63388bde62487e5c6b0ef2141a2beb5e9d"} Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.941711 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" event={"ID":"7d0ff181-18dc-47c5-b173-d5197b923176","Type":"ContainerStarted","Data":"a805957122029af7ec6c323dde0171b833884103a78ee11009cd1a50ab967611"} Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.949333 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.950237 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.952766 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.956269 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zjpgj" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.958459 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.967076 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.978510 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 15 20:18:54 crc kubenswrapper[4735]: I0215 20:18:54.998155 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.002543 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.009161 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.013001 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k64kb"] Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.020193 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.037700 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z6lf\" (UniqueName: \"kubernetes.io/projected/96ca6b03-36ef-47ec-9145-5b45994312cf-kube-api-access-2z6lf\") pod \"packageserver-d55dfcdfc-hm5c9\" (UID: \"96ca6b03-36ef-47ec-9145-5b45994312cf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.038605 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:18:55 crc kubenswrapper[4735]: W0215 20:18:55.043816 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode74e438d_61ef_403a_b958_8704f195f0ce.slice/crio-921be920b01c813ad90e0ba7b7ace4b4ccb8c46aed15fb8de3061bb4f4de21e0 WatchSource:0}: Error finding container 921be920b01c813ad90e0ba7b7ace4b4ccb8c46aed15fb8de3061bb4f4de21e0: Status 404 returned error can't find the container with id 921be920b01c813ad90e0ba7b7ace4b4ccb8c46aed15fb8de3061bb4f4de21e0 Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.054013 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x445g\" (UniqueName: \"kubernetes.io/projected/d3be498a-3362-4c71-85ac-e53b4889eec5-kube-api-access-x445g\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.054104 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc45t\" (UniqueName: \"kubernetes.io/projected/34482250-9472-45fe-aab5-485c7f129bca-kube-api-access-nc45t\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.059139 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jqvl\" (UniqueName: \"kubernetes.io/projected/ab54789a-45c9-43d4-a7ff-ad8e47f05af7-kube-api-access-2jqvl\") pod \"apiserver-7bbb656c7d-fg9hk\" (UID: \"ab54789a-45c9-43d4-a7ff-ad8e47f05af7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.078809 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f17ca838-95f0-4516-acf0-4d896c5be737-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-thglp\" (UID: \"f17ca838-95f0-4516-acf0-4d896c5be737\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.105910 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85tdz\" (UniqueName: \"kubernetes.io/projected/82f68827-538c-421d-ab03-2e0834d22429-kube-api-access-85tdz\") pod \"service-ca-operator-777779d784-dg5fw\" (UID: \"82f68827-538c-421d-ab03-2e0834d22429\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.129992 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjm5b\" (UniqueName: \"kubernetes.io/projected/a930872e-2679-46c9-ac6f-22bffe1b8d66-kube-api-access-hjm5b\") pod \"collect-profiles-29519775-4dm4h\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.148833 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmv2g\" (UniqueName: \"kubernetes.io/projected/61e54f93-d45e-47f0-b610-acd60abbf88c-kube-api-access-fmv2g\") pod \"service-ca-9c57cc56f-vspww\" (UID: \"61e54f93-d45e-47f0-b610-acd60abbf88c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.172921 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnfj6\" (UniqueName: \"kubernetes.io/projected/7ad4e51a-8091-4085-8c59-2a15dcc9a0ec-kube-api-access-fnfj6\") pod \"machine-config-controller-84d6567774-2jsgc\" (UID: \"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.197143 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh2gq\" (UniqueName: \"kubernetes.io/projected/5ee7e312-8482-4281-b3f1-9a3e3ca76aa2-kube-api-access-jh2gq\") pod \"machine-config-operator-74547568cd-4bzdb\" (UID: \"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.214695 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dfcj\" (UniqueName: \"kubernetes.io/projected/1626b474-7e91-4cde-bab7-a4db753478dd-kube-api-access-9dfcj\") pod \"marketplace-operator-79b997595-527r8\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.222570 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.229894 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.230881 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.235293 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2g69\" (UniqueName: \"kubernetes.io/projected/00ada6c8-9f70-4509-ad61-d31fb216090b-kube-api-access-c2g69\") pod \"multus-admission-controller-857f4d67dd-dwhtv\" (UID: \"00ada6c8-9f70-4509-ad61-d31fb216090b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.244161 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q85b\" (UniqueName: \"kubernetes.io/projected/9ec3f4d1-68c8-4460-8802-bc4ac3c1014f-kube-api-access-7q85b\") pod \"catalog-operator-68c6474976-nhpdn\" (UID: \"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.246329 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.278437 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.278995 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vspww" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.285336 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.300906 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.302564 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x76q8\" (UniqueName: \"kubernetes.io/projected/5cd27419-02e2-4228-b9f6-e6727c69722b-kube-api-access-x76q8\") pod \"package-server-manager-789f6589d5-v7mzq\" (UID: \"5cd27419-02e2-4228-b9f6-e6727c69722b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.318276 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.325691 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.338545 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/34482250-9472-45fe-aab5-485c7f129bca-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.339754 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.345337 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc45t\" (UniqueName: \"kubernetes.io/projected/34482250-9472-45fe-aab5-485c7f129bca-kube-api-access-nc45t\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:55 crc kubenswrapper[4735]: E0215 20:18:55.346157 4735 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:55 crc kubenswrapper[4735]: E0215 20:18:55.346280 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-config podName:34482250-9472-45fe-aab5-485c7f129bca nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.346257674 +0000 UTC m=+144.212273297 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-config") pod "machine-api-operator-5694c8668f-rjhgs" (UID: "34482250-9472-45fe-aab5-485c7f129bca") : failed to sync configmap cache: timed out waiting for the condition Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.354752 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x445g\" (UniqueName: \"kubernetes.io/projected/d3be498a-3362-4c71-85ac-e53b4889eec5-kube-api-access-x445g\") pod \"apiserver-76f77b778f-jbnv7\" (UID: \"d3be498a-3362-4c71-85ac-e53b4889eec5\") " pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.357454 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.428264 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.434633 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.468264 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.483403 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-certificates\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.483924 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/61a1542f-60bd-4453-a7c0-670088772a52-srv-cert\") pod \"olm-operator-6b444d44fb-xc855\" (UID: \"61a1542f-60bd-4453-a7c0-670088772a52\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484243 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d699b610-3540-4383-9af6-257c6821c498-default-certificate\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484387 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d4b1b306-2224-4908-b9f6-484d76a77ee1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484437 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/78551fcb-32d0-4021-9db2-c2ef4223a340-etcd-service-ca\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484501 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/44ee4a48-c32c-4690-92ee-8220e3881aea-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484594 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484636 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3036a2ba-e736-416a-bcf1-684d246cd3f1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4wxcd\" (UID: \"3036a2ba-e736-416a-bcf1-684d246cd3f1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484698 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/61a1542f-60bd-4453-a7c0-670088772a52-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xc855\" (UID: \"61a1542f-60bd-4453-a7c0-670088772a52\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484728 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78551fcb-32d0-4021-9db2-c2ef4223a340-serving-cert\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484801 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl2gn\" (UniqueName: \"kubernetes.io/projected/61a1542f-60bd-4453-a7c0-670088772a52-kube-api-access-jl2gn\") pod \"olm-operator-6b444d44fb-xc855\" (UID: \"61a1542f-60bd-4453-a7c0-670088772a52\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484840 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dbc306c-2daa-4f4c-89f7-bd8d1418b00b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xntbx\" (UID: \"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484870 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk64j\" (UniqueName: \"kubernetes.io/projected/3036a2ba-e736-416a-bcf1-684d246cd3f1-kube-api-access-dk64j\") pod \"control-plane-machine-set-operator-78cbb6b69f-4wxcd\" (UID: \"3036a2ba-e736-416a-bcf1-684d246cd3f1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484896 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d699b610-3540-4383-9af6-257c6821c498-metrics-certs\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484925 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-trusted-ca\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.484998 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n5x9\" (UniqueName: \"kubernetes.io/projected/d699b610-3540-4383-9af6-257c6821c498-kube-api-access-4n5x9\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485043 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9dbc306c-2daa-4f4c-89f7-bd8d1418b00b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xntbx\" (UID: \"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485098 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtjzh\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-kube-api-access-rtjzh\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485175 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9dbc306c-2daa-4f4c-89f7-bd8d1418b00b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xntbx\" (UID: \"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485202 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44ee4a48-c32c-4690-92ee-8220e3881aea-trusted-ca\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485236 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvvkw\" (UniqueName: \"kubernetes.io/projected/44ee4a48-c32c-4690-92ee-8220e3881aea-kube-api-access-xvvkw\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485265 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78551fcb-32d0-4021-9db2-c2ef4223a340-config\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485294 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf4nc\" (UniqueName: \"kubernetes.io/projected/78551fcb-32d0-4021-9db2-c2ef4223a340-kube-api-access-lf4nc\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485491 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/78551fcb-32d0-4021-9db2-c2ef4223a340-etcd-ca\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485580 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-bound-sa-token\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485604 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ctfq\" (UniqueName: \"kubernetes.io/projected/3195eae4-5944-4bb2-bd8e-d582a7181420-kube-api-access-9ctfq\") pod \"kube-storage-version-migrator-operator-b67b599dd-rbjqg\" (UID: \"3195eae4-5944-4bb2-bd8e-d582a7181420\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485636 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6rfv\" (UniqueName: \"kubernetes.io/projected/6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b-kube-api-access-g6rfv\") pod \"dns-operator-744455d44c-bnwh2\" (UID: \"6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485690 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3195eae4-5944-4bb2-bd8e-d582a7181420-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rbjqg\" (UID: \"3195eae4-5944-4bb2-bd8e-d582a7181420\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485733 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d4b1b306-2224-4908-b9f6-484d76a77ee1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485759 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b-metrics-tls\") pod \"dns-operator-744455d44c-bnwh2\" (UID: \"6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485791 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/44ee4a48-c32c-4690-92ee-8220e3881aea-metrics-tls\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485861 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-tls\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.485899 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d699b610-3540-4383-9af6-257c6821c498-stats-auth\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: E0215 20:18:55.488532 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:55.988511097 +0000 UTC m=+143.854526720 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.566659 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.568696 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3195eae4-5944-4bb2-bd8e-d582a7181420-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rbjqg\" (UID: \"3195eae4-5944-4bb2-bd8e-d582a7181420\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.568735 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/78551fcb-32d0-4021-9db2-c2ef4223a340-etcd-client\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.569984 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d699b610-3540-4383-9af6-257c6821c498-service-ca-bundle\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.574465 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.611237 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672045 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672257 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d699b610-3540-4383-9af6-257c6821c498-default-certificate\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672288 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d4b1b306-2224-4908-b9f6-484d76a77ee1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672317 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/78551fcb-32d0-4021-9db2-c2ef4223a340-etcd-service-ca\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672349 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a285b2fd-86ae-4f6f-af18-206a82cc17cf-config-volume\") pod \"dns-default-l8sf2\" (UID: \"a285b2fd-86ae-4f6f-af18-206a82cc17cf\") " pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672379 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/44ee4a48-c32c-4690-92ee-8220e3881aea-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672443 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns5rm\" (UniqueName: \"kubernetes.io/projected/c3962a27-53aa-4604-bc3d-e173e1525ab1-kube-api-access-ns5rm\") pod \"ingress-canary-2hhv9\" (UID: \"c3962a27-53aa-4604-bc3d-e173e1525ab1\") " pod="openshift-ingress-canary/ingress-canary-2hhv9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672480 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpvlg\" (UniqueName: \"kubernetes.io/projected/b2b33d6f-b149-404c-896c-b8b289180688-kube-api-access-xpvlg\") pod \"machine-config-server-f29nb\" (UID: \"b2b33d6f-b149-404c-896c-b8b289180688\") " pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672511 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3036a2ba-e736-416a-bcf1-684d246cd3f1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4wxcd\" (UID: \"3036a2ba-e736-416a-bcf1-684d246cd3f1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672567 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/61a1542f-60bd-4453-a7c0-670088772a52-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xc855\" (UID: \"61a1542f-60bd-4453-a7c0-670088772a52\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672582 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78551fcb-32d0-4021-9db2-c2ef4223a340-serving-cert\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-socket-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672657 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl2gn\" (UniqueName: \"kubernetes.io/projected/61a1542f-60bd-4453-a7c0-670088772a52-kube-api-access-jl2gn\") pod \"olm-operator-6b444d44fb-xc855\" (UID: \"61a1542f-60bd-4453-a7c0-670088772a52\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672681 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk64j\" (UniqueName: \"kubernetes.io/projected/3036a2ba-e736-416a-bcf1-684d246cd3f1-kube-api-access-dk64j\") pod \"control-plane-machine-set-operator-78cbb6b69f-4wxcd\" (UID: \"3036a2ba-e736-416a-bcf1-684d246cd3f1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672705 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d699b610-3540-4383-9af6-257c6821c498-metrics-certs\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672723 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dbc306c-2daa-4f4c-89f7-bd8d1418b00b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xntbx\" (UID: \"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672741 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-trusted-ca\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n5x9\" (UniqueName: \"kubernetes.io/projected/d699b610-3540-4383-9af6-257c6821c498-kube-api-access-4n5x9\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672809 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a285b2fd-86ae-4f6f-af18-206a82cc17cf-metrics-tls\") pod \"dns-default-l8sf2\" (UID: \"a285b2fd-86ae-4f6f-af18-206a82cc17cf\") " pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672843 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9dbc306c-2daa-4f4c-89f7-bd8d1418b00b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xntbx\" (UID: \"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672876 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtjzh\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-kube-api-access-rtjzh\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672917 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9dbc306c-2daa-4f4c-89f7-bd8d1418b00b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xntbx\" (UID: \"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672934 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44ee4a48-c32c-4690-92ee-8220e3881aea-trusted-ca\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672967 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvvkw\" (UniqueName: \"kubernetes.io/projected/44ee4a48-c32c-4690-92ee-8220e3881aea-kube-api-access-xvvkw\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.672983 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf4nc\" (UniqueName: \"kubernetes.io/projected/78551fcb-32d0-4021-9db2-c2ef4223a340-kube-api-access-lf4nc\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673001 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-csi-data-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673029 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78551fcb-32d0-4021-9db2-c2ef4223a340-config\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673051 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b2b33d6f-b149-404c-896c-b8b289180688-certs\") pod \"machine-config-server-f29nb\" (UID: \"b2b33d6f-b149-404c-896c-b8b289180688\") " pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673088 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/78551fcb-32d0-4021-9db2-c2ef4223a340-etcd-ca\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673162 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-bound-sa-token\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673180 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ctfq\" (UniqueName: \"kubernetes.io/projected/3195eae4-5944-4bb2-bd8e-d582a7181420-kube-api-access-9ctfq\") pod \"kube-storage-version-migrator-operator-b67b599dd-rbjqg\" (UID: \"3195eae4-5944-4bb2-bd8e-d582a7181420\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673198 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3962a27-53aa-4604-bc3d-e173e1525ab1-cert\") pod \"ingress-canary-2hhv9\" (UID: \"c3962a27-53aa-4604-bc3d-e173e1525ab1\") " pod="openshift-ingress-canary/ingress-canary-2hhv9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673216 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6rfv\" (UniqueName: \"kubernetes.io/projected/6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b-kube-api-access-g6rfv\") pod \"dns-operator-744455d44c-bnwh2\" (UID: \"6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673231 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-registration-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673279 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3195eae4-5944-4bb2-bd8e-d582a7181420-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rbjqg\" (UID: \"3195eae4-5944-4bb2-bd8e-d582a7181420\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673295 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjrfj\" (UniqueName: \"kubernetes.io/projected/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-kube-api-access-wjrfj\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673312 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d4b1b306-2224-4908-b9f6-484d76a77ee1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673329 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b-metrics-tls\") pod \"dns-operator-744455d44c-bnwh2\" (UID: \"6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673356 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/44ee4a48-c32c-4690-92ee-8220e3881aea-metrics-tls\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673471 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbsw5\" (UniqueName: \"kubernetes.io/projected/a285b2fd-86ae-4f6f-af18-206a82cc17cf-kube-api-access-kbsw5\") pod \"dns-default-l8sf2\" (UID: \"a285b2fd-86ae-4f6f-af18-206a82cc17cf\") " pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673537 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-tls\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673614 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d699b610-3540-4383-9af6-257c6821c498-stats-auth\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673707 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3195eae4-5944-4bb2-bd8e-d582a7181420-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rbjqg\" (UID: \"3195eae4-5944-4bb2-bd8e-d582a7181420\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673727 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/78551fcb-32d0-4021-9db2-c2ef4223a340-etcd-client\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673794 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d699b610-3540-4383-9af6-257c6821c498-service-ca-bundle\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673826 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-mountpoint-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673874 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-certificates\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673930 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b2b33d6f-b149-404c-896c-b8b289180688-node-bootstrap-token\") pod \"machine-config-server-f29nb\" (UID: \"b2b33d6f-b149-404c-896c-b8b289180688\") " pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.673998 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/61a1542f-60bd-4453-a7c0-670088772a52-srv-cert\") pod \"olm-operator-6b444d44fb-xc855\" (UID: \"61a1542f-60bd-4453-a7c0-670088772a52\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.674015 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-plugins-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.674809 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/78551fcb-32d0-4021-9db2-c2ef4223a340-etcd-ca\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: E0215 20:18:55.677350 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.177330328 +0000 UTC m=+144.043345951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.686009 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/78551fcb-32d0-4021-9db2-c2ef4223a340-etcd-service-ca\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.686665 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d4b1b306-2224-4908-b9f6-484d76a77ee1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.699374 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3195eae4-5944-4bb2-bd8e-d582a7181420-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rbjqg\" (UID: \"3195eae4-5944-4bb2-bd8e-d582a7181420\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.701340 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9dbc306c-2daa-4f4c-89f7-bd8d1418b00b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xntbx\" (UID: \"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.703603 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44ee4a48-c32c-4690-92ee-8220e3881aea-trusted-ca\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.706161 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d699b610-3540-4383-9af6-257c6821c498-default-certificate\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.723742 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3195eae4-5944-4bb2-bd8e-d582a7181420-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rbjqg\" (UID: \"3195eae4-5944-4bb2-bd8e-d582a7181420\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.724868 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d4b1b306-2224-4908-b9f6-484d76a77ee1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.725583 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dbc306c-2daa-4f4c-89f7-bd8d1418b00b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xntbx\" (UID: \"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.726581 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-certificates\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.727809 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d699b610-3540-4383-9af6-257c6821c498-service-ca-bundle\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.732722 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b-metrics-tls\") pod \"dns-operator-744455d44c-bnwh2\" (UID: \"6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.740760 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-trusted-ca\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.741762 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78551fcb-32d0-4021-9db2-c2ef4223a340-config\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.743334 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d699b610-3540-4383-9af6-257c6821c498-metrics-certs\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.747112 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/44ee4a48-c32c-4690-92ee-8220e3881aea-metrics-tls\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.774855 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/61a1542f-60bd-4453-a7c0-670088772a52-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xc855\" (UID: \"61a1542f-60bd-4453-a7c0-670088772a52\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.785401 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78551fcb-32d0-4021-9db2-c2ef4223a340-serving-cert\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.786928 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a285b2fd-86ae-4f6f-af18-206a82cc17cf-metrics-tls\") pod \"dns-default-l8sf2\" (UID: \"a285b2fd-86ae-4f6f-af18-206a82cc17cf\") " pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797254 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-csi-data-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797324 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b2b33d6f-b149-404c-896c-b8b289180688-certs\") pod \"machine-config-server-f29nb\" (UID: \"b2b33d6f-b149-404c-896c-b8b289180688\") " pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797373 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3962a27-53aa-4604-bc3d-e173e1525ab1-cert\") pod \"ingress-canary-2hhv9\" (UID: \"c3962a27-53aa-4604-bc3d-e173e1525ab1\") " pod="openshift-ingress-canary/ingress-canary-2hhv9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797396 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-registration-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797416 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjrfj\" (UniqueName: \"kubernetes.io/projected/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-kube-api-access-wjrfj\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797450 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbsw5\" (UniqueName: \"kubernetes.io/projected/a285b2fd-86ae-4f6f-af18-206a82cc17cf-kube-api-access-kbsw5\") pod \"dns-default-l8sf2\" (UID: \"a285b2fd-86ae-4f6f-af18-206a82cc17cf\") " pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797524 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-mountpoint-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797547 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b2b33d6f-b149-404c-896c-b8b289180688-node-bootstrap-token\") pod \"machine-config-server-f29nb\" (UID: \"b2b33d6f-b149-404c-896c-b8b289180688\") " pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797568 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-plugins-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797612 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a285b2fd-86ae-4f6f-af18-206a82cc17cf-config-volume\") pod \"dns-default-l8sf2\" (UID: \"a285b2fd-86ae-4f6f-af18-206a82cc17cf\") " pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797642 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns5rm\" (UniqueName: \"kubernetes.io/projected/c3962a27-53aa-4604-bc3d-e173e1525ab1-kube-api-access-ns5rm\") pod \"ingress-canary-2hhv9\" (UID: \"c3962a27-53aa-4604-bc3d-e173e1525ab1\") " pod="openshift-ingress-canary/ingress-canary-2hhv9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797667 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797686 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpvlg\" (UniqueName: \"kubernetes.io/projected/b2b33d6f-b149-404c-896c-b8b289180688-kube-api-access-xpvlg\") pod \"machine-config-server-f29nb\" (UID: \"b2b33d6f-b149-404c-896c-b8b289180688\") " pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.797735 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-socket-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.798728 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-socket-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.798820 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-csi-data-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.803507 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-plugins-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.804142 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a285b2fd-86ae-4f6f-af18-206a82cc17cf-config-volume\") pod \"dns-default-l8sf2\" (UID: \"a285b2fd-86ae-4f6f-af18-206a82cc17cf\") " pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:55 crc kubenswrapper[4735]: E0215 20:18:55.804457 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.304442905 +0000 UTC m=+144.170458528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.806271 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-registration-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.821575 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-mountpoint-dir\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.833655 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/78551fcb-32d0-4021-9db2-c2ef4223a340-etcd-client\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.833788 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a285b2fd-86ae-4f6f-af18-206a82cc17cf-metrics-tls\") pod \"dns-default-l8sf2\" (UID: \"a285b2fd-86ae-4f6f-af18-206a82cc17cf\") " pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.845738 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9dbc306c-2daa-4f4c-89f7-bd8d1418b00b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xntbx\" (UID: \"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.854562 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl2gn\" (UniqueName: \"kubernetes.io/projected/61a1542f-60bd-4453-a7c0-670088772a52-kube-api-access-jl2gn\") pod \"olm-operator-6b444d44fb-xc855\" (UID: \"61a1542f-60bd-4453-a7c0-670088772a52\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.855295 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3036a2ba-e736-416a-bcf1-684d246cd3f1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4wxcd\" (UID: \"3036a2ba-e736-416a-bcf1-684d246cd3f1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.855728 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/61a1542f-60bd-4453-a7c0-670088772a52-srv-cert\") pod \"olm-operator-6b444d44fb-xc855\" (UID: \"61a1542f-60bd-4453-a7c0-670088772a52\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.857118 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-tls\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.857290 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d699b610-3540-4383-9af6-257c6821c498-stats-auth\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.866406 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk64j\" (UniqueName: \"kubernetes.io/projected/3036a2ba-e736-416a-bcf1-684d246cd3f1-kube-api-access-dk64j\") pod \"control-plane-machine-set-operator-78cbb6b69f-4wxcd\" (UID: \"3036a2ba-e736-416a-bcf1-684d246cd3f1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.867397 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.873501 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf4nc\" (UniqueName: \"kubernetes.io/projected/78551fcb-32d0-4021-9db2-c2ef4223a340-kube-api-access-lf4nc\") pod \"etcd-operator-b45778765-2w25g\" (UID: \"78551fcb-32d0-4021-9db2-c2ef4223a340\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.889267 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n5x9\" (UniqueName: \"kubernetes.io/projected/d699b610-3540-4383-9af6-257c6821c498-kube-api-access-4n5x9\") pod \"router-default-5444994796-lj984\" (UID: \"d699b610-3540-4383-9af6-257c6821c498\") " pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.889934 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b2b33d6f-b149-404c-896c-b8b289180688-node-bootstrap-token\") pod \"machine-config-server-f29nb\" (UID: \"b2b33d6f-b149-404c-896c-b8b289180688\") " pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.893228 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3962a27-53aa-4604-bc3d-e173e1525ab1-cert\") pod \"ingress-canary-2hhv9\" (UID: \"c3962a27-53aa-4604-bc3d-e173e1525ab1\") " pod="openshift-ingress-canary/ingress-canary-2hhv9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.893608 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b2b33d6f-b149-404c-896c-b8b289180688-certs\") pod \"machine-config-server-f29nb\" (UID: \"b2b33d6f-b149-404c-896c-b8b289180688\") " pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.893862 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.894624 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ctfq\" (UniqueName: \"kubernetes.io/projected/3195eae4-5944-4bb2-bd8e-d582a7181420-kube-api-access-9ctfq\") pod \"kube-storage-version-migrator-operator-b67b599dd-rbjqg\" (UID: \"3195eae4-5944-4bb2-bd8e-d582a7181420\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.898555 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:55 crc kubenswrapper[4735]: E0215 20:18:55.898669 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.398649525 +0000 UTC m=+144.264665148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.899169 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-bound-sa-token\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.907938 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/44ee4a48-c32c-4690-92ee-8220e3881aea-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.908519 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtjzh\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-kube-api-access-rtjzh\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.914134 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjrfj\" (UniqueName: \"kubernetes.io/projected/fc8fe06e-5292-4b4c-9b20-b2a3f4fce798-kube-api-access-wjrfj\") pod \"csi-hostpathplugin-lkzd9\" (UID: \"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798\") " pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.918180 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvvkw\" (UniqueName: \"kubernetes.io/projected/44ee4a48-c32c-4690-92ee-8220e3881aea-kube-api-access-xvvkw\") pod \"ingress-operator-5b745b69d9-ws9hh\" (UID: \"44ee4a48-c32c-4690-92ee-8220e3881aea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.941335 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns5rm\" (UniqueName: \"kubernetes.io/projected/c3962a27-53aa-4604-bc3d-e173e1525ab1-kube-api-access-ns5rm\") pod \"ingress-canary-2hhv9\" (UID: \"c3962a27-53aa-4604-bc3d-e173e1525ab1\") " pod="openshift-ingress-canary/ingress-canary-2hhv9" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.941898 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.950820 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6rfv\" (UniqueName: \"kubernetes.io/projected/6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b-kube-api-access-g6rfv\") pod \"dns-operator-744455d44c-bnwh2\" (UID: \"6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.955223 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpvlg\" (UniqueName: \"kubernetes.io/projected/b2b33d6f-b149-404c-896c-b8b289180688-kube-api-access-xpvlg\") pod \"machine-config-server-f29nb\" (UID: \"b2b33d6f-b149-404c-896c-b8b289180688\") " pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.973689 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:55 crc kubenswrapper[4735]: E0215 20:18:55.974162 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.474147457 +0000 UTC m=+144.340163080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.991711 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" Feb 15 20:18:55 crc kubenswrapper[4735]: I0215 20:18:55.991740 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq"] Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.013594 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.030675 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zjpgj"] Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.037219 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd"] Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.052590 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.065606 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbsw5\" (UniqueName: \"kubernetes.io/projected/a285b2fd-86ae-4f6f-af18-206a82cc17cf-kube-api-access-kbsw5\") pod \"dns-default-l8sf2\" (UID: \"a285b2fd-86ae-4f6f-af18-206a82cc17cf\") " pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.075816 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k64kb" event={"ID":"e74e438d-61ef-403a-b958-8704f195f0ce","Type":"ContainerStarted","Data":"9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853"} Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.075853 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k64kb" event={"ID":"e74e438d-61ef-403a-b958-8704f195f0ce","Type":"ContainerStarted","Data":"921be920b01c813ad90e0ba7b7ace4b4ccb8c46aed15fb8de3061bb4f4de21e0"} Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.077552 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.087387 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-x4gh2"] Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.114922 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.614888165 +0000 UTC m=+144.480903788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.116875 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.182479 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.185212 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.685195158 +0000 UTC m=+144.551210781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.209235 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2hhv9" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.218906 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-l8sf2" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.226403 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.233698 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nz76" podStartSLOduration=122.233664263 podStartE2EDuration="2m2.233664263s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:56.225513248 +0000 UTC m=+144.091528871" watchObservedRunningTime="2026-02-15 20:18:56.233664263 +0000 UTC m=+144.099679886" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.250786 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-f29nb" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.251406 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.283564 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp"] Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.292658 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.307018 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.806880999 +0000 UTC m=+144.672896622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.308198 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkgwz" podStartSLOduration=122.308184026 podStartE2EDuration="2m2.308184026s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:56.298422196 +0000 UTC m=+144.164437819" watchObservedRunningTime="2026-02-15 20:18:56.308184026 +0000 UTC m=+144.174199649" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.348968 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8htb6" podStartSLOduration=122.348929319 podStartE2EDuration="2m2.348929319s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:56.345411618 +0000 UTC m=+144.211427241" watchObservedRunningTime="2026-02-15 20:18:56.348929319 +0000 UTC m=+144.214944942" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.349098 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9"] Feb 15 20:18:56 crc kubenswrapper[4735]: W0215 20:18:56.393548 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5afc2ea_9954_4617_bd56_5493fc53121f.slice/crio-4432d7d17156664538292bcde13fac372196df1dc731ee175c124751b47e0c0d WatchSource:0}: Error finding container 4432d7d17156664538292bcde13fac372196df1dc731ee175c124751b47e0c0d: Status 404 returned error can't find the container with id 4432d7d17156664538292bcde13fac372196df1dc731ee175c124751b47e0c0d Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.395785 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-config\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.395824 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.396416 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.896400234 +0000 UTC m=+144.762415857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.397033 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34482250-9472-45fe-aab5-485c7f129bca-config\") pod \"machine-api-operator-5694c8668f-rjhgs\" (UID: \"34482250-9472-45fe-aab5-485c7f129bca\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.462061 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2"] Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.487142 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.499080 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.499579 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:56.999543422 +0000 UTC m=+144.865559035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.564984 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" podStartSLOduration=122.564963933 podStartE2EDuration="2m2.564963933s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:56.518112545 +0000 UTC m=+144.384128188" watchObservedRunningTime="2026-02-15 20:18:56.564963933 +0000 UTC m=+144.430979556" Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.582380 4735 csr.go:261] certificate signing request csr-5ld45 is approved, waiting to be issued Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.606404 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.606978 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.106958811 +0000 UTC m=+144.972974434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.614672 4735 csr.go:257] certificate signing request csr-5ld45 is issued Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.707635 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.708114 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.20808189 +0000 UTC m=+145.074097503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.708298 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.708641 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.208627146 +0000 UTC m=+145.074642759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.770427 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vspww"] Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.819981 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.820450 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.320431492 +0000 UTC m=+145.186447115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.857110 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw"] Feb 15 20:18:56 crc kubenswrapper[4735]: I0215 20:18:56.922673 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:56 crc kubenswrapper[4735]: E0215 20:18:56.923063 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.423050294 +0000 UTC m=+145.289065917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:56 crc kubenswrapper[4735]: W0215 20:18:56.973934 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd699b610_3540_4383_9af6_257c6821c498.slice/crio-3eb9bdbad06083ffdb37b5208bc892f957113dddb8411b251beacefb7a0f2b21 WatchSource:0}: Error finding container 3eb9bdbad06083ffdb37b5208bc892f957113dddb8411b251beacefb7a0f2b21: Status 404 returned error can't find the container with id 3eb9bdbad06083ffdb37b5208bc892f957113dddb8411b251beacefb7a0f2b21 Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.010386 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp"] Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.024604 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.025878 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.525860231 +0000 UTC m=+145.391875854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.130207 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.131017 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.631000566 +0000 UTC m=+145.497016189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.133025 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h"] Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.232077 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" event={"ID":"96ca6b03-36ef-47ec-9145-5b45994312cf","Type":"ContainerStarted","Data":"54f797025f1473454102f0d11493cbefd782f81c3e88e68dfde65760e44f35bc"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.233426 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.233793 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.733774683 +0000 UTC m=+145.599790306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.324582 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr8hd" podStartSLOduration=123.324553234 podStartE2EDuration="2m3.324553234s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:57.3226793 +0000 UTC m=+145.188694923" watchObservedRunningTime="2026-02-15 20:18:57.324553234 +0000 UTC m=+145.190568857" Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.337751 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.339031 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.83901757 +0000 UTC m=+145.705033193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.385152 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2" event={"ID":"1d593c77-f754-4464-bc94-5c265029d330","Type":"ContainerStarted","Data":"5007ecd431d262bf44ddd68615d288781a41dc3ad08472b4afc26b1ca8784fc0"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.387817 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fvv22" podStartSLOduration=123.387798184 podStartE2EDuration="2m3.387798184s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:57.38665164 +0000 UTC m=+145.252667263" watchObservedRunningTime="2026-02-15 20:18:57.387798184 +0000 UTC m=+145.253813807" Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.440386 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.443242 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-f29nb" event={"ID":"b2b33d6f-b149-404c-896c-b8b289180688","Type":"ContainerStarted","Data":"791d80997841e8c956ae70d079f213c7e8b00f3a28209e2425132ac6d0d863f0"} Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.443420 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.943397273 +0000 UTC m=+145.809412896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.444872 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.445417 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:57.945403031 +0000 UTC m=+145.811418654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.460861 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" podStartSLOduration=123.460824655 podStartE2EDuration="2m3.460824655s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:57.459358712 +0000 UTC m=+145.325374335" watchObservedRunningTime="2026-02-15 20:18:57.460824655 +0000 UTC m=+145.326840278" Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.461593 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" event={"ID":"82f68827-538c-421d-ab03-2e0834d22429","Type":"ContainerStarted","Data":"91c7350049c3ec5dbd55f4bdfc9c20c4afe8236d859a885c8afd7177c7a04598"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.475095 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zjpgj" event={"ID":"0faea847-63f9-488b-9f3a-080e36fd9025","Type":"ContainerStarted","Data":"405264c58b9a639cc7a283f1b0b0a700b28f70ef433db4812c7a6edea28f1510"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.478529 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-zjpgj" Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.489886 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-zjpgj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.490051 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zjpgj" podUID="0faea847-63f9-488b-9f3a-080e36fd9025" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.507173 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-x4gh2" event={"ID":"20355c26-2666-4915-87d3-b61747ed5692","Type":"ContainerStarted","Data":"a4bea401e1df006c6275788dbc64c8feb54bddb08a53bfb713a23f0692640600"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.507866 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.512991 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vspww" event={"ID":"61e54f93-d45e-47f0-b610-acd60abbf88c","Type":"ContainerStarted","Data":"a20231a653c77557548c765e5fb586cdc5418383709dc17ee4e1dd5f4fab5af0"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.534751 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" podStartSLOduration=123.53473448 podStartE2EDuration="2m3.53473448s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:57.533124495 +0000 UTC m=+145.399140118" watchObservedRunningTime="2026-02-15 20:18:57.53473448 +0000 UTC m=+145.400750103" Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.536721 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" event={"ID":"d5afc2ea-9954-4617-bd56-5493fc53121f","Type":"ContainerStarted","Data":"4432d7d17156664538292bcde13fac372196df1dc731ee175c124751b47e0c0d"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.540630 4735 patch_prober.go:28] interesting pod/console-operator-58897d9998-x4gh2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.540668 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-x4gh2" podUID="20355c26-2666-4915-87d3-b61747ed5692" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.554035 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.554717 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:58.054680074 +0000 UTC m=+145.920695687 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.571825 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" event={"ID":"3359cd7b-e283-4f9c-baf6-29e6e85e247e","Type":"ContainerStarted","Data":"7aeca5c1ce06180d0f4f8e2e686be707f4613bd2b3bddc6583c6f6bd87b161d4"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.585462 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lj984" event={"ID":"d699b610-3540-4383-9af6-257c6821c498","Type":"ContainerStarted","Data":"3eb9bdbad06083ffdb37b5208bc892f957113dddb8411b251beacefb7a0f2b21"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.604152 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" event={"ID":"f17ca838-95f0-4516-acf0-4d896c5be737","Type":"ContainerStarted","Data":"3af211a5dfda8e7def218e7f03db0410d8ff6926582d4b347cc8d7dc4519b704"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.623149 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-15 20:13:56 +0000 UTC, rotation deadline is 2026-11-18 20:17:47.236323274 +0000 UTC Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.623184 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6623h58m49.613141469s for next certificate rotation Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.638098 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" event={"ID":"27821f2f-df7e-42d2-b687-244c7b6e5a3f","Type":"ContainerStarted","Data":"d3eba9ea918633b59cceabd72822d6ac8c4c081f8d8bdda6972775f37126e073"} Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.659075 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.660764 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:58.160748765 +0000 UTC m=+146.026764378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.762492 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.764022 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:58.263992266 +0000 UTC m=+146.130007889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.868789 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.869475 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:58.36946194 +0000 UTC m=+146.235477563 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:57 crc kubenswrapper[4735]: I0215 20:18:57.969799 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:57 crc kubenswrapper[4735]: E0215 20:18:57.970317 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:58.47027402 +0000 UTC m=+146.336289643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.007803 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.074260 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:58 crc kubenswrapper[4735]: E0215 20:18:58.075312 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:58.575290721 +0000 UTC m=+146.441306344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.176291 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:58 crc kubenswrapper[4735]: E0215 20:18:58.176663 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:58.676644486 +0000 UTC m=+146.542660109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.219209 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2w25g"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.292648 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-527r8"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.308572 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:58 crc kubenswrapper[4735]: E0215 20:18:58.309145 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:58.809031244 +0000 UTC m=+146.675046867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: W0215 20:18:58.343294 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1626b474_7e91_4cde_bab7_a4db753478dd.slice/crio-331dd034b95df0dea652eec75c06b10d23cb27166085827d1292ea96443e3366 WatchSource:0}: Error finding container 331dd034b95df0dea652eec75c06b10d23cb27166085827d1292ea96443e3366: Status 404 returned error can't find the container with id 331dd034b95df0dea652eec75c06b10d23cb27166085827d1292ea96443e3366 Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.351746 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dwhtv"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.398510 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-zjpgj" podStartSLOduration=124.398485378 podStartE2EDuration="2m4.398485378s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:58.35159169 +0000 UTC m=+146.217607313" watchObservedRunningTime="2026-02-15 20:18:58.398485378 +0000 UTC m=+146.264501001" Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.403986 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.407086 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-x4gh2" podStartSLOduration=124.407061674 podStartE2EDuration="2m4.407061674s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:58.394579856 +0000 UTC m=+146.260595479" watchObservedRunningTime="2026-02-15 20:18:58.407061674 +0000 UTC m=+146.273077297" Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.409607 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:58 crc kubenswrapper[4735]: E0215 20:18:58.409934 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:58.909915846 +0000 UTC m=+146.775931459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.435083 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jbnv7"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.484262 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-k64kb" podStartSLOduration=124.484241785 podStartE2EDuration="2m4.484241785s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:58.449331781 +0000 UTC m=+146.315347404" watchObservedRunningTime="2026-02-15 20:18:58.484241785 +0000 UTC m=+146.350257408" Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.485846 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.521627 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:58 crc kubenswrapper[4735]: E0215 20:18:58.523550 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.023536515 +0000 UTC m=+146.889552138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.594157 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.631934 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:58 crc kubenswrapper[4735]: E0215 20:18:58.632319 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.132302725 +0000 UTC m=+146.998318348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.685343 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.697645 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" event={"ID":"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f","Type":"ContainerStarted","Data":"81ce6a4c6d4cd717094936f60201d1fc3656c5b53950e292518c53a64da559ec"} Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.697696 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" event={"ID":"9ec3f4d1-68c8-4460-8802-bc4ac3c1014f","Type":"ContainerStarted","Data":"52fd5f238d1fe5136cd9beb6986d9a163f47d1c47097bddcb4144bb1fecda5be"} Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.698094 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.703440 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2hhv9"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.717436 4735 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-nhpdn container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.717478 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" podUID="9ec3f4d1-68c8-4460-8802-bc4ac3c1014f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.729243 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lj984" event={"ID":"d699b610-3540-4383-9af6-257c6821c498","Type":"ContainerStarted","Data":"4ab7b4e29cdf87ab6df424866433f3db6ad732ce347a395b0f87ae13c0aa50f4"} Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.733625 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:58 crc kubenswrapper[4735]: E0215 20:18:58.734008 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.233992279 +0000 UTC m=+147.100007902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.738204 4735 generic.go:334] "Generic (PLEG): container finished" podID="27821f2f-df7e-42d2-b687-244c7b6e5a3f" containerID="083ef3ab571aa3b5084eaa636faed14b773098071c141b16a3975772fcf0dd0c" exitCode=0 Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.738896 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" event={"ID":"27821f2f-df7e-42d2-b687-244c7b6e5a3f","Type":"ContainerDied","Data":"083ef3ab571aa3b5084eaa636faed14b773098071c141b16a3975772fcf0dd0c"} Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.765110 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.798181 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" podStartSLOduration=124.798161886 podStartE2EDuration="2m4.798161886s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:58.796683714 +0000 UTC m=+146.662699337" watchObservedRunningTime="2026-02-15 20:18:58.798161886 +0000 UTC m=+146.664177509" Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.820199 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-l8sf2"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.820513 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" event={"ID":"f17ca838-95f0-4516-acf0-4d896c5be737","Type":"ContainerStarted","Data":"a9393efd6d841b4b8da4b75fc2570aa31ddc361ab89f794127e88d1d0d975255"} Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.834746 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:58 crc kubenswrapper[4735]: E0215 20:18:58.836362 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.336331094 +0000 UTC m=+147.202346707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.870356 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.870412 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bnwh2"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.880043 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.896464 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2" event={"ID":"1d593c77-f754-4464-bc94-5c265029d330","Type":"ContainerStarted","Data":"551c974ec5679ce3c3862bc8f9591692bf83573a70460ad7c12c0db5ba0bdfc3"} Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.896522 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2" event={"ID":"1d593c77-f754-4464-bc94-5c265029d330","Type":"ContainerStarted","Data":"b5218ca6be3610d1274dc21cee163c44edea6f8ecc8fd8f9f7edc889e2c76966"} Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.897335 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lj984" podStartSLOduration=124.897308397 podStartE2EDuration="2m4.897308397s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:58.892134659 +0000 UTC m=+146.758150292" watchObservedRunningTime="2026-02-15 20:18:58.897308397 +0000 UTC m=+146.763324010" Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.941911 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.941978 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" event={"ID":"96ca6b03-36ef-47ec-9145-5b45994312cf","Type":"ContainerStarted","Data":"f4033871747b6d58402d3daf9a61b2a6a86593d4c553d6eb2407daab53128ead"} Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.956647 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:58 crc kubenswrapper[4735]: E0215 20:18:58.959005 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.458981191 +0000 UTC m=+147.324996814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.987828 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855"] Feb 15 20:18:58 crc kubenswrapper[4735]: I0215 20:18:58.988028 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" event={"ID":"d3be498a-3362-4c71-85ac-e53b4889eec5","Type":"ContainerStarted","Data":"ded130ea0d5c3f3cc21f998ea34bcbae04b3201f4264174c3c36b0d3718a425d"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.016968 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-x4gh2" event={"ID":"20355c26-2666-4915-87d3-b61747ed5692","Type":"ContainerStarted","Data":"d711cb97784a0ecff0a4d13b242a3e48b8b619d4e6991875fedb453660cc5420"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.018448 4735 patch_prober.go:28] interesting pod/console-operator-58897d9998-x4gh2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.018530 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-x4gh2" podUID="20355c26-2666-4915-87d3-b61747ed5692" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.026061 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rjhgs"] Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.036997 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9s6h2" podStartSLOduration=125.036971275 podStartE2EDuration="2m5.036971275s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.000053903 +0000 UTC m=+146.866069536" watchObservedRunningTime="2026-02-15 20:18:59.036971275 +0000 UTC m=+146.902986898" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.054933 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.057861 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.057911 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.057970 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.058622 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.558602588 +0000 UTC m=+147.424618211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.061235 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lkzd9"] Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.062104 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-thglp" podStartSLOduration=125.062091448 podStartE2EDuration="2m5.062091448s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.044922984 +0000 UTC m=+146.910938617" watchObservedRunningTime="2026-02-15 20:18:59.062091448 +0000 UTC m=+146.928107071" Feb 15 20:18:59 crc kubenswrapper[4735]: W0215 20:18:59.070634 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34482250_9472_45fe_aab5_485c7f129bca.slice/crio-998a85bef29b5942c332b749362601780a5d744d23c4006df234c13cd4fb2eb6 WatchSource:0}: Error finding container 998a85bef29b5942c332b749362601780a5d744d23c4006df234c13cd4fb2eb6: Status 404 returned error can't find the container with id 998a85bef29b5942c332b749362601780a5d744d23c4006df234c13cd4fb2eb6 Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.071064 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" event={"ID":"a930872e-2679-46c9-ac6f-22bffe1b8d66","Type":"ContainerStarted","Data":"fd2feb5ae65c846e3bfc78616e103db5f08dd6d485bc8e98f0fea784efdca889"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.071099 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" event={"ID":"a930872e-2679-46c9-ac6f-22bffe1b8d66","Type":"ContainerStarted","Data":"262a806652391feb5cb48d58492781291011665125e16c3339dca4a80856a9bc"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.073477 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vspww" event={"ID":"61e54f93-d45e-47f0-b610-acd60abbf88c","Type":"ContainerStarted","Data":"fa5d2663fef635eaacc80d88c17e643e9fd41b31f43dd34b6568903ef0e61822"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.094967 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" podStartSLOduration=125.094932023 podStartE2EDuration="2m5.094932023s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.094381467 +0000 UTC m=+146.960397090" watchObservedRunningTime="2026-02-15 20:18:59.094932023 +0000 UTC m=+146.960947646" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.108487 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" event={"ID":"d5afc2ea-9954-4617-bd56-5493fc53121f","Type":"ContainerStarted","Data":"e30e82d10a8267bad3b4b143beb6d1ee024606ea21fca26dc7035b8f055e6428"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.128554 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" event={"ID":"82f68827-538c-421d-ab03-2e0834d22429","Type":"ContainerStarted","Data":"112939fd42d0a348fe2646ad16c3f3dbad4b4f333d90950bbf27fd0d3c0f8c8e"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.130367 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" podStartSLOduration=125.130323431 podStartE2EDuration="2m5.130323431s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.118382767 +0000 UTC m=+146.984398390" watchObservedRunningTime="2026-02-15 20:18:59.130323431 +0000 UTC m=+146.996339054" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.165512 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-vspww" podStartSLOduration=125.165493982 podStartE2EDuration="2m5.165493982s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.165094291 +0000 UTC m=+147.031109914" watchObservedRunningTime="2026-02-15 20:18:59.165493982 +0000 UTC m=+147.031509605" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.165667 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.174471 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.67445366 +0000 UTC m=+147.540469283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.203598 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zjpgj" event={"ID":"0faea847-63f9-488b-9f3a-080e36fd9025","Type":"ContainerStarted","Data":"48f8b7f77f6b6d09672afbcb1f9abb302e0441889686cbfb7afde17460b6bd34"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.205311 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-zjpgj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.205362 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zjpgj" podUID="0faea847-63f9-488b-9f3a-080e36fd9025" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.208639 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dg5fw" podStartSLOduration=125.208615584 podStartE2EDuration="2m5.208615584s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.203195808 +0000 UTC m=+147.069211441" watchObservedRunningTime="2026-02-15 20:18:59.208615584 +0000 UTC m=+147.074631207" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.245336 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" event={"ID":"78551fcb-32d0-4021-9db2-c2ef4223a340","Type":"ContainerStarted","Data":"2e1a8318c2d9ad1bdfaf433d231cc81b07ae76845005dc784efa3151b61ea908"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.246663 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qjxgp" podStartSLOduration=125.246639257 podStartE2EDuration="2m5.246639257s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.245591807 +0000 UTC m=+147.111607430" watchObservedRunningTime="2026-02-15 20:18:59.246639257 +0000 UTC m=+147.112654880" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.266385 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.267358 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" event={"ID":"3359cd7b-e283-4f9c-baf6-29e6e85e247e","Type":"ContainerStarted","Data":"a022c11629f3f5afbef48a5772a088e4b5936d8a80f9255627e5f209be923109"} Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.281754 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.781720896 +0000 UTC m=+147.647736519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.281909 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.282374 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.782367225 +0000 UTC m=+147.648382848 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.308603 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gd7bq" podStartSLOduration=125.308584759 podStartE2EDuration="2m5.308584759s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.30689181 +0000 UTC m=+147.172907433" watchObservedRunningTime="2026-02-15 20:18:59.308584759 +0000 UTC m=+147.174600382" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.311712 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-f29nb" event={"ID":"b2b33d6f-b149-404c-896c-b8b289180688","Type":"ContainerStarted","Data":"48355787e3443464b232d9cfd01ed7de524262b10e0149db25a57a311ab34876"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.336260 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" event={"ID":"00ada6c8-9f70-4509-ad61-d31fb216090b","Type":"ContainerStarted","Data":"afd5902598e508a729abc4a0435d3332934d04bf8317a096830386ad9f42e716"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.342824 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-f29nb" podStartSLOduration=7.342796693 podStartE2EDuration="7.342796693s" podCreationTimestamp="2026-02-15 20:18:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.342648488 +0000 UTC m=+147.208664131" watchObservedRunningTime="2026-02-15 20:18:59.342796693 +0000 UTC m=+147.208812306" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.342862 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" event={"ID":"5cd27419-02e2-4228-b9f6-e6727c69722b","Type":"ContainerStarted","Data":"3fbff585d03499fa3af98a26e963ec8acd2bb801247ecd86082581fa208b5212"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.382199 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" event={"ID":"1626b474-7e91-4cde-bab7-a4db753478dd","Type":"ContainerStarted","Data":"331dd034b95df0dea652eec75c06b10d23cb27166085827d1292ea96443e3366"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.383213 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.384640 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.385160 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.885137752 +0000 UTC m=+147.751153365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.407251 4735 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-527r8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.407318 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.410363 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" event={"ID":"3036a2ba-e736-416a-bcf1-684d246cd3f1","Type":"ContainerStarted","Data":"902a910aca825d7f2383b612aee2753b108a9d1e6d9665e7a507dd495c1228cb"} Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.459413 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" podStartSLOduration=125.459391607 podStartE2EDuration="2m5.459391607s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:18:59.45843898 +0000 UTC m=+147.324454603" watchObservedRunningTime="2026-02-15 20:18:59.459391607 +0000 UTC m=+147.325407230" Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.490576 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.492548 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:18:59.992530651 +0000 UTC m=+147.858546274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.591510 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.592478 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.092442055 +0000 UTC m=+147.958457678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.693668 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.694183 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.194166751 +0000 UTC m=+148.060182374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.797565 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.798289 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.298258145 +0000 UTC m=+148.164273768 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.798487 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.798862 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.298853912 +0000 UTC m=+148.164869525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:18:59 crc kubenswrapper[4735]: I0215 20:18:59.898986 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:18:59 crc kubenswrapper[4735]: E0215 20:18:59.899278 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.399260361 +0000 UTC m=+148.265275984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.012002 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.012343 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.512328473 +0000 UTC m=+148.378344096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.066323 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:00 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:00 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:00 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.066786 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.118536 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.119029 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.619006793 +0000 UTC m=+148.485022416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.220768 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.221172 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.721157351 +0000 UTC m=+148.587172974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.323309 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.323524 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.823493375 +0000 UTC m=+148.689508998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.324770 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.325192 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.825175723 +0000 UTC m=+148.691191346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.426978 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.427429 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:00.927413645 +0000 UTC m=+148.793429268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.501569 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" event={"ID":"34482250-9472-45fe-aab5-485c7f129bca","Type":"ContainerStarted","Data":"7c3586b5e4f44ddbbc2e285dba1723fe2165dc2a3301dcbe8fdf91f418777b87"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.501626 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" event={"ID":"34482250-9472-45fe-aab5-485c7f129bca","Type":"ContainerStarted","Data":"998a85bef29b5942c332b749362601780a5d744d23c4006df234c13cd4fb2eb6"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.529751 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.530149 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.030132129 +0000 UTC m=+148.896147752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.575798 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" event={"ID":"1626b474-7e91-4cde-bab7-a4db753478dd","Type":"ContainerStarted","Data":"be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.577682 4735 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-527r8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.577747 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.630424 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.631009 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.13097691 +0000 UTC m=+148.996992533 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.632718 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" event={"ID":"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2","Type":"ContainerStarted","Data":"d567de4885dedb6397ad412ac335f142bee3d022f78f09ef602bd95c991cee9d"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.632796 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" event={"ID":"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2","Type":"ContainerStarted","Data":"1d2c0058f497387f91b0c284d92a6d727806e2921369542300cde7fc987cce64"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.662896 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" event={"ID":"6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b","Type":"ContainerStarted","Data":"418ba92fa354896a7ff8403321231480ea2ba5507b762807b40ebb8f509428b8"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.680307 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" event={"ID":"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b","Type":"ContainerStarted","Data":"9f667f87ef7b556d77b3d7fd5e696c035a32bd5b2bfaed99e710b46f6481d242"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.719134 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" event={"ID":"5cd27419-02e2-4228-b9f6-e6727c69722b","Type":"ContainerStarted","Data":"42a65e1a6306b3dfd2726383499d00b127d2c59629a4a460b873dac12767f702"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.732324 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.732699 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.232685036 +0000 UTC m=+149.098700659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.756905 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" event={"ID":"44ee4a48-c32c-4690-92ee-8220e3881aea","Type":"ContainerStarted","Data":"3288e06a72e4b60ba3df4720223bf6c1ddfc237e018f67c804ff09cb30e31459"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.756997 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" event={"ID":"44ee4a48-c32c-4690-92ee-8220e3881aea","Type":"ContainerStarted","Data":"8f3c130f7a3b61b7c01949f25954e4eb74470d24637923aa02089ce181a02bcb"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.780550 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2hhv9" event={"ID":"c3962a27-53aa-4604-bc3d-e173e1525ab1","Type":"ContainerStarted","Data":"cafa8ba542baae1f01ec4914ab5c60d9319f1ebf510b40cac10b9e4fbb2384f9"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.780628 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2hhv9" event={"ID":"c3962a27-53aa-4604-bc3d-e173e1525ab1","Type":"ContainerStarted","Data":"7467494eb24b5dec973b198711d313bf7aad58bc4f51146f229f58963a7798df"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.784459 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" podStartSLOduration=126.784447185 podStartE2EDuration="2m6.784447185s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:00.782825419 +0000 UTC m=+148.648841042" watchObservedRunningTime="2026-02-15 20:19:00.784447185 +0000 UTC m=+148.650462808" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.829371 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2hhv9" podStartSLOduration=8.829352497 podStartE2EDuration="8.829352497s" podCreationTimestamp="2026-02-15 20:18:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:00.828634456 +0000 UTC m=+148.694650069" watchObservedRunningTime="2026-02-15 20:19:00.829352497 +0000 UTC m=+148.695368110" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.834590 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" event={"ID":"27821f2f-df7e-42d2-b687-244c7b6e5a3f","Type":"ContainerStarted","Data":"ffb4fc27f4136a7eaccc25fe244481a3132c68fd463f777199d29d9bd3fe32ae"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.835006 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.835328 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.835544 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.335518794 +0000 UTC m=+149.201534407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.835693 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.836007 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.336000789 +0000 UTC m=+149.202016412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.843570 4735 generic.go:334] "Generic (PLEG): container finished" podID="d3be498a-3362-4c71-85ac-e53b4889eec5" containerID="7af4b8ddcbccaa512a524680970eabb9207e99aa9e7d05169d1def482623e9ca" exitCode=0 Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.843806 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" event={"ID":"d3be498a-3362-4c71-85ac-e53b4889eec5","Type":"ContainerDied","Data":"7af4b8ddcbccaa512a524680970eabb9207e99aa9e7d05169d1def482623e9ca"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.848685 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" event={"ID":"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798","Type":"ContainerStarted","Data":"9d6436ecf2ef7e49ca828d695f65bd8b72cad4c0b15537fc8089d8c5103fc618"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.886668 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" event={"ID":"00ada6c8-9f70-4509-ad61-d31fb216090b","Type":"ContainerStarted","Data":"b31ae36fb6b6dd24db4e8d90bfbb46fb72e53e57e916a3c6c6d76b20edc53d47"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.888784 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" event={"ID":"ab54789a-45c9-43d4-a7ff-ad8e47f05af7","Type":"ContainerStarted","Data":"6a3476c54fb752e025606bbe8d3c7e05257e6fd9bade9be69bdf55ca6ff3114d"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.921146 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" event={"ID":"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec","Type":"ContainerStarted","Data":"0240981f7f23eab78bbbf322d3032296085fcfdfdf8c7668a06b60ce5bf68553"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.921188 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" event={"ID":"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec","Type":"ContainerStarted","Data":"ddf8d73107f814e4bf43c40d0185b8fbc08b77cfe71f67d77f5374190d9393a7"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.930265 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-l8sf2" event={"ID":"a285b2fd-86ae-4f6f-af18-206a82cc17cf","Type":"ContainerStarted","Data":"ca9d593f7d52dcb75b795c0cacccad9fdfab8517cd9d7341f12501a68e2ec576"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.930321 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-l8sf2" event={"ID":"a285b2fd-86ae-4f6f-af18-206a82cc17cf","Type":"ContainerStarted","Data":"3cca30fb2700b3a8225371096b0eb549c2e84dbda433e2f86f84b772cdba8045"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.937569 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.937748 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.437718754 +0000 UTC m=+149.303734387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.937855 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.937921 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.938209 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.938322 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:19:00 crc kubenswrapper[4735]: E0215 20:19:00.940242 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.440222857 +0000 UTC m=+149.306238480 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.945060 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.953715 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" podStartSLOduration=126.953690304 podStartE2EDuration="2m6.953690304s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:00.908229786 +0000 UTC m=+148.774245429" watchObservedRunningTime="2026-02-15 20:19:00.953690304 +0000 UTC m=+148.819705927" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.955601 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" event={"ID":"3195eae4-5944-4bb2-bd8e-d582a7181420","Type":"ContainerStarted","Data":"07d4063c2bf4a733f01558477e1d810571345d0898fe988955ac2994e9693d5d"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.955639 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" event={"ID":"3195eae4-5944-4bb2-bd8e-d582a7181420","Type":"ContainerStarted","Data":"f38e4583921245f5c45f0078fa4e200a3711f6bf92b0ca2696d38a528b85a569"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.984484 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.989231 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" event={"ID":"3036a2ba-e736-416a-bcf1-684d246cd3f1","Type":"ContainerStarted","Data":"e2f377df032d4f7a372dc60d53f29dc351a2652986c270eef33a8f4b4739bccb"} Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.990541 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:19:00 crc kubenswrapper[4735]: I0215 20:19:00.999519 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" podStartSLOduration=126.999501642 podStartE2EDuration="2m6.999501642s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:00.955843925 +0000 UTC m=+148.821859548" watchObservedRunningTime="2026-02-15 20:19:00.999501642 +0000 UTC m=+148.865517265" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.000401 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" event={"ID":"78551fcb-32d0-4021-9db2-c2ef4223a340","Type":"ContainerStarted","Data":"db1911ee43b97631bcdb15041bb166d7e8b4d45ad0005e843f201790c560ef86"} Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.026142 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.027654 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" event={"ID":"61a1542f-60bd-4453-a7c0-670088772a52","Type":"ContainerStarted","Data":"23a15b91714daad4795e5cd14e744676caa87d94542732333c528db9ab060028"} Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.027712 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.027727 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" event={"ID":"61a1542f-60bd-4453-a7c0-670088772a52","Type":"ContainerStarted","Data":"4ddfc2bb4700b7d20438374febd6689af414785b559504d8106804931db90f50"} Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.037509 4735 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xc855 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.037562 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" podUID="61a1542f-60bd-4453-a7c0-670088772a52" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.039581 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.040035 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.040823 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.54080227 +0000 UTC m=+149.406817883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.065635 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-zjpgj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.065706 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zjpgj" podUID="0faea847-63f9-488b-9f3a-080e36fd9025" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.076706 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.076780 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.090478 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rbjqg" podStartSLOduration=127.090458028 podStartE2EDuration="2m7.090458028s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:01.029484855 +0000 UTC m=+148.895500478" watchObservedRunningTime="2026-02-15 20:19:01.090458028 +0000 UTC m=+148.956473651" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.091472 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:01 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:01 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:01 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.098625 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.101151 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.160656 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.163701 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4wxcd" podStartSLOduration=127.163679674 podStartE2EDuration="2m7.163679674s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:01.0943306 +0000 UTC m=+148.960346223" watchObservedRunningTime="2026-02-15 20:19:01.163679674 +0000 UTC m=+149.029695297" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.164339 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-2w25g" podStartSLOduration=127.164333124 podStartE2EDuration="2m7.164333124s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:01.152392939 +0000 UTC m=+149.018408562" watchObservedRunningTime="2026-02-15 20:19:01.164333124 +0000 UTC m=+149.030348747" Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.166812 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.666784124 +0000 UTC m=+149.532799747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.215367 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.216698 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" podStartSLOduration=127.216686499 podStartE2EDuration="2m7.216686499s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:01.216293928 +0000 UTC m=+149.082309551" watchObservedRunningTime="2026-02-15 20:19:01.216686499 +0000 UTC m=+149.082702122" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.265221 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.265866 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.765848464 +0000 UTC m=+149.631864077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.302396 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-x4gh2" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.348244 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.369805 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.370184 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.870170565 +0000 UTC m=+149.736186188 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.470746 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.471116 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:01.971080247 +0000 UTC m=+149.837095860 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.573153 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.573528 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.073515725 +0000 UTC m=+149.939531338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.674130 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.674502 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.174484119 +0000 UTC m=+150.040499742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.775760 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.776384 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.27636999 +0000 UTC m=+150.142385613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.888715 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.889043 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.38900808 +0000 UTC m=+150.255023693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.889475 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.889929 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.389911726 +0000 UTC m=+150.255927359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:01 crc kubenswrapper[4735]: I0215 20:19:01.991596 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:01 crc kubenswrapper[4735]: E0215 20:19:01.992058 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.492037174 +0000 UTC m=+150.358052797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.028668 4735 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-hm5c9 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.028744 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" podUID="96ca6b03-36ef-47ec-9145-5b45994312cf" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.070699 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:02 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:02 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:02 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.070785 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.089606 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" event={"ID":"00ada6c8-9f70-4509-ad61-d31fb216090b","Type":"ContainerStarted","Data":"7b1679061fa1bb5e37f2ff39e47c3100c83db4ed5a7748b5449b4acdf8678b1d"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.094203 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:02 crc kubenswrapper[4735]: E0215 20:19:02.094685 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.594669207 +0000 UTC m=+150.460684830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.105292 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" event={"ID":"5cd27419-02e2-4228-b9f6-e6727c69722b","Type":"ContainerStarted","Data":"27c0dfe4314af2f7b6187edcd8eb3998d7b793866e64446902945024e0c68aa3"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.106070 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.129104 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-dwhtv" podStartSLOduration=128.128881881 podStartE2EDuration="2m8.128881881s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:02.127739588 +0000 UTC m=+149.993755211" watchObservedRunningTime="2026-02-15 20:19:02.128881881 +0000 UTC m=+149.994897504" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.141879 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" event={"ID":"5ee7e312-8482-4281-b3f1-9a3e3ca76aa2","Type":"ContainerStarted","Data":"94c870ae5a48c88be82e51126d13cf133ae4b9301d3440caf6137c72a7895376"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.159242 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2jsgc" event={"ID":"7ad4e51a-8091-4085-8c59-2a15dcc9a0ec","Type":"ContainerStarted","Data":"e972b5eedd432b786e345dfe697cf2abbd5b8a91e21afdc09cc330ed0584bd70"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.167925 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" podStartSLOduration=128.167908053 podStartE2EDuration="2m8.167908053s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:02.16537226 +0000 UTC m=+150.031387883" watchObservedRunningTime="2026-02-15 20:19:02.167908053 +0000 UTC m=+150.033923676" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.169340 4735 generic.go:334] "Generic (PLEG): container finished" podID="ab54789a-45c9-43d4-a7ff-ad8e47f05af7" containerID="683a85773050a71217dc3378f54bf25efc80c6ff71df450200b392ba8ad2aab5" exitCode=0 Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.169484 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" event={"ID":"ab54789a-45c9-43d4-a7ff-ad8e47f05af7","Type":"ContainerDied","Data":"683a85773050a71217dc3378f54bf25efc80c6ff71df450200b392ba8ad2aab5"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.196825 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" event={"ID":"34482250-9472-45fe-aab5-485c7f129bca","Type":"ContainerStarted","Data":"fdcda888845023b71b33cbb44e43adec1fb2f541233b71eb9e2c319f7fe096e9"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.197359 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:02 crc kubenswrapper[4735]: E0215 20:19:02.198231 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.698194775 +0000 UTC m=+150.564210398 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.250213 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" event={"ID":"d3be498a-3362-4c71-85ac-e53b4889eec5","Type":"ContainerStarted","Data":"97d4da07b89be7dee83a047225957c5f9040bf5b2d99e2836e1e3a8a3f0d35f7"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.276413 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-rjhgs" podStartSLOduration=128.276397105 podStartE2EDuration="2m8.276397105s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:02.275531949 +0000 UTC m=+150.141547572" watchObservedRunningTime="2026-02-15 20:19:02.276397105 +0000 UTC m=+150.142412728" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.276510 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4bzdb" podStartSLOduration=128.276507158 podStartE2EDuration="2m8.276507158s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:02.217928742 +0000 UTC m=+150.083944365" watchObservedRunningTime="2026-02-15 20:19:02.276507158 +0000 UTC m=+150.142522781" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.284238 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" event={"ID":"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798","Type":"ContainerStarted","Data":"363e495ea6c002619868759e5c463c7c5461cbffece1c3c8b36cb5acd3d3ad43"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.299215 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:02 crc kubenswrapper[4735]: E0215 20:19:02.299557 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.79954282 +0000 UTC m=+150.665558433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.323531 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-l8sf2" event={"ID":"a285b2fd-86ae-4f6f-af18-206a82cc17cf","Type":"ContainerStarted","Data":"442f0b22c3aa9b0f8bb204b279b7c53b85389c71374d03148896579ed7f60a24"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.324316 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-l8sf2" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.338163 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ws9hh" event={"ID":"44ee4a48-c32c-4690-92ee-8220e3881aea","Type":"ContainerStarted","Data":"b159927741d3f640e7bbcca8f8d45216127fa8f506a5bdde973935920e8aab14"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.370516 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" event={"ID":"9dbc306c-2daa-4f4c-89f7-bd8d1418b00b","Type":"ContainerStarted","Data":"76c5c0bee864544ae4adc5d65476d2b6c02661e51dae110c94cded510969f2c1"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.373374 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" event={"ID":"6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b","Type":"ContainerStarted","Data":"5f30ed872f2945c66ba4781dcf4e5a9777f594d142e582b98d373c35c1a785fc"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.373394 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" event={"ID":"6c432c1e-c5b1-4f0f-85c6-6daff4b3ab0b","Type":"ContainerStarted","Data":"f261529d37053265970833ad1fcb3ca882f4883f4ebf1c5b55974f2c43db72f0"} Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.375863 4735 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-527r8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.375928 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.403624 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:02 crc kubenswrapper[4735]: E0215 20:19:02.405390 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:02.905351064 +0000 UTC m=+150.771366737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.441503 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xc855" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.506027 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:02 crc kubenswrapper[4735]: E0215 20:19:02.506938 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.006921956 +0000 UTC m=+150.872937579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.513399 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-l8sf2" podStartSLOduration=10.513371231 podStartE2EDuration="10.513371231s" podCreationTimestamp="2026-02-15 20:18:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:02.433433022 +0000 UTC m=+150.299448655" watchObservedRunningTime="2026-02-15 20:19:02.513371231 +0000 UTC m=+150.379386854" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.513997 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-bnwh2" podStartSLOduration=128.513991429 podStartE2EDuration="2m8.513991429s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:02.507686578 +0000 UTC m=+150.373702201" watchObservedRunningTime="2026-02-15 20:19:02.513991429 +0000 UTC m=+150.380007052" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.556048 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z2cm8"] Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.558089 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.568961 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.572663 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xntbx" podStartSLOduration=128.572633446 podStartE2EDuration="2m8.572633446s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:02.568910809 +0000 UTC m=+150.434926432" watchObservedRunningTime="2026-02-15 20:19:02.572633446 +0000 UTC m=+150.438649069" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.599095 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2cm8"] Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.607905 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:02 crc kubenswrapper[4735]: E0215 20:19:02.608857 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.108835527 +0000 UTC m=+150.974851150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.710831 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-utilities\") pod \"certified-operators-z2cm8\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.710875 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.710898 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-catalog-content\") pod \"certified-operators-z2cm8\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.710955 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxxxw\" (UniqueName: \"kubernetes.io/projected/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-kube-api-access-mxxxw\") pod \"certified-operators-z2cm8\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: E0215 20:19:02.711297 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.211281015 +0000 UTC m=+151.077296638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.751086 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lzl2l"] Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.752070 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.769364 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.813898 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.814104 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-utilities\") pod \"certified-operators-z2cm8\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.814141 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-catalog-content\") pod \"certified-operators-z2cm8\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.814186 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxxxw\" (UniqueName: \"kubernetes.io/projected/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-kube-api-access-mxxxw\") pod \"certified-operators-z2cm8\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: E0215 20:19:02.814586 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.314570866 +0000 UTC m=+151.180586489 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.834755 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-utilities\") pod \"certified-operators-z2cm8\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.836160 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-catalog-content\") pod \"certified-operators-z2cm8\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.839444 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lzl2l"] Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.886365 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxxxw\" (UniqueName: \"kubernetes.io/projected/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-kube-api-access-mxxxw\") pod \"certified-operators-z2cm8\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.915700 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-utilities\") pod \"community-operators-lzl2l\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.915744 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-catalog-content\") pod \"community-operators-lzl2l\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.915766 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsvrk\" (UniqueName: \"kubernetes.io/projected/31a015ae-51a2-488b-8bfd-4829857d6c7d-kube-api-access-wsvrk\") pod \"community-operators-lzl2l\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.915830 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:02 crc kubenswrapper[4735]: E0215 20:19:02.916245 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.41623064 +0000 UTC m=+151.282246263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.956100 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:19:02 crc kubenswrapper[4735]: W0215 20:19:02.959611 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-08da43fe935b7fc1950bfb5315fddc0adc11977123c7df78014d7bcfeb964e0a WatchSource:0}: Error finding container 08da43fe935b7fc1950bfb5315fddc0adc11977123c7df78014d7bcfeb964e0a: Status 404 returned error can't find the container with id 08da43fe935b7fc1950bfb5315fddc0adc11977123c7df78014d7bcfeb964e0a Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.988734 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7ml7r"] Feb 15 20:19:02 crc kubenswrapper[4735]: I0215 20:19:02.989694 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.019937 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.020176 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-utilities\") pod \"community-operators-lzl2l\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.020206 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-catalog-content\") pod \"community-operators-lzl2l\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.020224 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsvrk\" (UniqueName: \"kubernetes.io/projected/31a015ae-51a2-488b-8bfd-4829857d6c7d-kube-api-access-wsvrk\") pod \"community-operators-lzl2l\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.020623 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.520607333 +0000 UTC m=+151.386622956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.021014 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-utilities\") pod \"community-operators-lzl2l\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.021217 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-catalog-content\") pod \"community-operators-lzl2l\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.053754 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7ml7r"] Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.073321 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:03 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:03 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:03 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.073396 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.100511 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hm5c9" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.122557 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-catalog-content\") pod \"certified-operators-7ml7r\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.122607 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-utilities\") pod \"certified-operators-7ml7r\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.122660 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.122686 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnlcx\" (UniqueName: \"kubernetes.io/projected/bf815092-d21d-4b8b-bbfa-fd11550ffb26-kube-api-access-vnlcx\") pod \"certified-operators-7ml7r\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.123706 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.623688578 +0000 UTC m=+151.489704201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.129698 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsvrk\" (UniqueName: \"kubernetes.io/projected/31a015ae-51a2-488b-8bfd-4829857d6c7d-kube-api-access-wsvrk\") pod \"community-operators-lzl2l\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.136025 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.162790 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f9ft4"] Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.174486 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.225541 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.226263 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-catalog-content\") pod \"certified-operators-7ml7r\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.226298 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-utilities\") pod \"certified-operators-7ml7r\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.226365 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnlcx\" (UniqueName: \"kubernetes.io/projected/bf815092-d21d-4b8b-bbfa-fd11550ffb26-kube-api-access-vnlcx\") pod \"certified-operators-7ml7r\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.226768 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.726750233 +0000 UTC m=+151.592765856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.227196 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-catalog-content\") pod \"certified-operators-7ml7r\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.227399 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-utilities\") pod \"certified-operators-7ml7r\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.250747 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f9ft4"] Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.302262 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnlcx\" (UniqueName: \"kubernetes.io/projected/bf815092-d21d-4b8b-bbfa-fd11550ffb26-kube-api-access-vnlcx\") pod \"certified-operators-7ml7r\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.326052 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.327223 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-catalog-content\") pod \"community-operators-f9ft4\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.327260 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27lfn\" (UniqueName: \"kubernetes.io/projected/4f5c12d8-7152-4299-bcad-c59201575285-kube-api-access-27lfn\") pod \"community-operators-f9ft4\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.327279 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-utilities\") pod \"community-operators-f9ft4\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.327325 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.327624 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.827610285 +0000 UTC m=+151.693625898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.393046 4735 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qmtqd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": context deadline exceeded" start-of-body= Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.393131 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" podUID="27821f2f-df7e-42d2-b687-244c7b6e5a3f" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": context deadline exceeded" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.429792 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.429924 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.929898087 +0000 UTC m=+151.795913700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.430803 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.431033 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-catalog-content\") pod \"community-operators-f9ft4\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.431118 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27lfn\" (UniqueName: \"kubernetes.io/projected/4f5c12d8-7152-4299-bcad-c59201575285-kube-api-access-27lfn\") pod \"community-operators-f9ft4\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.431145 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-utilities\") pod \"community-operators-f9ft4\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.431484 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:03.931460962 +0000 UTC m=+151.797476585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.431899 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-utilities\") pod \"community-operators-f9ft4\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.432492 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-catalog-content\") pod \"community-operators-f9ft4\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.444607 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6e4ea72742f2a53b5198e7efdf245113910c28834e109b118d65713d6260dcc2"} Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.444699 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6db60d37fe57e283c7096883cf2add02162d0bb617f338acc725de9f4412029d"} Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.495983 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" event={"ID":"d3be498a-3362-4c71-85ac-e53b4889eec5","Type":"ContainerStarted","Data":"309435d6212730d6a29025f880934dd18c31428cadb80ea4e79f213075ce361c"} Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.534395 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7a036c00064b6da71a166e26abf055c66da6b78ddc2e22c327d86e425b6a9a70"} Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.534457 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"da8c878c703642db1c54f57c34e7477d415826ff3116cbe94f3f9cd172b32eaa"} Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.546606 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.548146 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.048122998 +0000 UTC m=+151.914138621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.559329 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"08da43fe935b7fc1950bfb5315fddc0adc11977123c7df78014d7bcfeb964e0a"} Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.593176 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" event={"ID":"ab54789a-45c9-43d4-a7ff-ad8e47f05af7","Type":"ContainerStarted","Data":"61cc95ca93e3a807abf077964075ace8c30c76bc62f255ee08b80d34de9b74a8"} Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.652794 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.654823 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.154804657 +0000 UTC m=+152.020820270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.657838 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27lfn\" (UniqueName: \"kubernetes.io/projected/4f5c12d8-7152-4299-bcad-c59201575285-kube-api-access-27lfn\") pod \"community-operators-f9ft4\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.762803 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.768654 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.268634681 +0000 UTC m=+152.134650294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.771120 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.771458 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.271450062 +0000 UTC m=+152.137465675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.818896 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.874676 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.874771 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.374742964 +0000 UTC m=+152.240758587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.875319 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.875868 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.375852756 +0000 UTC m=+152.241868379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.925692 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" podStartSLOduration=129.925671059 podStartE2EDuration="2m9.925671059s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:03.920553012 +0000 UTC m=+151.786568635" watchObservedRunningTime="2026-02-15 20:19:03.925671059 +0000 UTC m=+151.791686682" Feb 15 20:19:03 crc kubenswrapper[4735]: I0215 20:19:03.976629 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:03 crc kubenswrapper[4735]: E0215 20:19:03.978688 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.478666693 +0000 UTC m=+152.344682316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.038038 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lzl2l"] Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.078108 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:04 crc kubenswrapper[4735]: E0215 20:19:04.078492 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.578475935 +0000 UTC m=+152.444491558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.093519 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:04 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:04 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:04 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.093587 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.181903 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:04 crc kubenswrapper[4735]: E0215 20:19:04.182352 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.682334192 +0000 UTC m=+152.548349815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.284264 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:04 crc kubenswrapper[4735]: E0215 20:19:04.291498 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.791478752 +0000 UTC m=+152.657494375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.295620 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" podStartSLOduration=130.29557954 podStartE2EDuration="2m10.29557954s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:04.277252043 +0000 UTC m=+152.143267686" watchObservedRunningTime="2026-02-15 20:19:04.29557954 +0000 UTC m=+152.161595163" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.385913 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:04 crc kubenswrapper[4735]: E0215 20:19:04.386327 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.88630907 +0000 UTC m=+152.752324693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.460895 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.462284 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.478908 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.479188 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.481925 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.492603 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:04 crc kubenswrapper[4735]: E0215 20:19:04.494402 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:04.994386339 +0000 UTC m=+152.860401962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.545532 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7ml7r"] Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.565004 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hm2zs"] Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.579363 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.584045 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.593205 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.593572 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.593631 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.593777 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hm2zs"] Feb 15 20:19:04 crc kubenswrapper[4735]: E0215 20:19:04.594120 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.094087527 +0000 UTC m=+152.960103150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.665136 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzl2l" event={"ID":"31a015ae-51a2-488b-8bfd-4829857d6c7d","Type":"ContainerStarted","Data":"14a6772ae1869b63804bdcd96fb8e8336faf4f28f68af2c0cc165f7e405b4936"} Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.682941 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" event={"ID":"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798","Type":"ContainerStarted","Data":"58644717aef0148df6d924de38b88652cd0e54bccef5e2d2232501b09dc3c262"} Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.684324 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a03bff07341ffb4c61a16d231dbda160b9ac5839916eb721441d2ad4bb9e1ba6"} Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.685311 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.688338 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ml7r" event={"ID":"bf815092-d21d-4b8b-bbfa-fd11550ffb26","Type":"ContainerStarted","Data":"d26a1ac054ce4dd551726ca95029c8a1899acc5e4ffaa4c806e4caa0aba77e4b"} Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.697399 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.697449 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tgrt\" (UniqueName: \"kubernetes.io/projected/59b9213c-60ff-4ecd-99d9-4c1fe670592c-kube-api-access-2tgrt\") pod \"redhat-marketplace-hm2zs\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.697490 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.697518 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-catalog-content\") pod \"redhat-marketplace-hm2zs\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.697534 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-utilities\") pod \"redhat-marketplace-hm2zs\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.697560 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.697635 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:04 crc kubenswrapper[4735]: E0215 20:19:04.698240 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.198225323 +0000 UTC m=+153.064240946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.699850 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2cm8"] Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.760801 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.790494 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.791886 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.798588 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.798772 4735 patch_prober.go:28] interesting pod/console-f9d7485db-k64kb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.798825 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-k64kb" podUID="e74e438d-61ef-403a-b958-8704f195f0ce" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.798912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-utilities\") pod \"redhat-marketplace-hm2zs\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.799337 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tgrt\" (UniqueName: \"kubernetes.io/projected/59b9213c-60ff-4ecd-99d9-4c1fe670592c-kube-api-access-2tgrt\") pod \"redhat-marketplace-hm2zs\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.799537 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-catalog-content\") pod \"redhat-marketplace-hm2zs\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: E0215 20:19:04.800646 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.300628789 +0000 UTC m=+153.166644412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.801000 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-utilities\") pod \"redhat-marketplace-hm2zs\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.803975 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-catalog-content\") pod \"redhat-marketplace-hm2zs\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.832196 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.894863 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tgrt\" (UniqueName: \"kubernetes.io/projected/59b9213c-60ff-4ecd-99d9-4c1fe670592c-kube-api-access-2tgrt\") pod \"redhat-marketplace-hm2zs\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.901139 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:04 crc kubenswrapper[4735]: E0215 20:19:04.901570 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.401553393 +0000 UTC m=+153.267569016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.921559 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.958423 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.958462 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4bxj7"] Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.964058 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-zjpgj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.964109 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zjpgj" podUID="0faea847-63f9-488b-9f3a-080e36fd9025" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.964356 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-zjpgj container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.964397 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-zjpgj" podUID="0faea847-63f9-488b-9f3a-080e36fd9025" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.969897 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:04 crc kubenswrapper[4735]: I0215 20:19:04.981147 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bxj7"] Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.002472 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.002937 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.502919388 +0000 UTC m=+153.368935011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.064168 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:05 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:05 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:05 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.064223 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.104873 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-catalog-content\") pod \"redhat-marketplace-4bxj7\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.104960 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.105023 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-utilities\") pod \"redhat-marketplace-4bxj7\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.105077 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55xzs\" (UniqueName: \"kubernetes.io/projected/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-kube-api-access-55xzs\") pod \"redhat-marketplace-4bxj7\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.105387 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.605374376 +0000 UTC m=+153.471389999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.207205 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.207347 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.707318228 +0000 UTC m=+153.573333851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.207528 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-utilities\") pod \"redhat-marketplace-4bxj7\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.207587 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55xzs\" (UniqueName: \"kubernetes.io/projected/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-kube-api-access-55xzs\") pod \"redhat-marketplace-4bxj7\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.207617 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-catalog-content\") pod \"redhat-marketplace-4bxj7\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.207676 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.208049 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.708039489 +0000 UTC m=+153.574055112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.208587 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-utilities\") pod \"redhat-marketplace-4bxj7\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.209141 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-catalog-content\") pod \"redhat-marketplace-4bxj7\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.278859 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55xzs\" (UniqueName: \"kubernetes.io/projected/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-kube-api-access-55xzs\") pod \"redhat-marketplace-4bxj7\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.309476 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.309978 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.809954791 +0000 UTC m=+153.675970414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.328485 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.328523 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.336996 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.383843 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f9ft4"] Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.411737 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.412186 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:05.912172511 +0000 UTC m=+153.778188134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.521760 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.522707 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.0226694 +0000 UTC m=+153.888685023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.525852 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.531622 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.031568586 +0000 UTC m=+153.897584209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.561591 4735 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-fg9hk container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 15 20:19:05 crc kubenswrapper[4735]: [+]log ok Feb 15 20:19:05 crc kubenswrapper[4735]: [+]etcd ok Feb 15 20:19:05 crc kubenswrapper[4735]: [-]poststarthook/start-apiserver-admission-initializer failed: reason withheld Feb 15 20:19:05 crc kubenswrapper[4735]: [-]poststarthook/generic-apiserver-start-informers failed: reason withheld Feb 15 20:19:05 crc kubenswrapper[4735]: [+]poststarthook/max-in-flight-filter ok Feb 15 20:19:05 crc kubenswrapper[4735]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 15 20:19:05 crc kubenswrapper[4735]: [+]poststarthook/openshift.io-StartUserInformer ok Feb 15 20:19:05 crc kubenswrapper[4735]: [+]poststarthook/openshift.io-StartOAuthInformer ok Feb 15 20:19:05 crc kubenswrapper[4735]: [-]poststarthook/openshift.io-StartTokenTimeoutUpdater failed: reason withheld Feb 15 20:19:05 crc kubenswrapper[4735]: livez check failed Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.561731 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" podUID="ab54789a-45c9-43d4-a7ff-ad8e47f05af7" containerName="oauth-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.578083 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.612669 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.612781 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.634366 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.637208 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.137165173 +0000 UTC m=+154.003180796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.647164 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.647660 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.147642285 +0000 UTC m=+154.013657908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.753196 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.754596 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.254574791 +0000 UTC m=+154.120590414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.776787 4735 generic.go:334] "Generic (PLEG): container finished" podID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerID="8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801" exitCode=0 Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.777276 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2cm8" event={"ID":"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23","Type":"ContainerDied","Data":"8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801"} Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.777315 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2cm8" event={"ID":"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23","Type":"ContainerStarted","Data":"84e25a0254db6ef7dee717854d2c2820780f0346f7062c33e70662acce2a3c1a"} Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.787761 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.826072 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" event={"ID":"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798","Type":"ContainerStarted","Data":"32af5f3cfb28297617454e534609cb846c8febc66c1540806985c7403b53d289"} Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.839341 4735 generic.go:334] "Generic (PLEG): container finished" podID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerID="1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8" exitCode=0 Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.839555 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ml7r" event={"ID":"bf815092-d21d-4b8b-bbfa-fd11550ffb26","Type":"ContainerDied","Data":"1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8"} Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.849768 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9ft4" event={"ID":"4f5c12d8-7152-4299-bcad-c59201575285","Type":"ContainerStarted","Data":"2aa4b66863df797b15a8d564c7a7fbd345ca09ef4bda4c56a0490a4c008948f4"} Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.856260 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.856636 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.356615937 +0000 UTC m=+154.222631560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.881419 4735 generic.go:334] "Generic (PLEG): container finished" podID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerID="9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba" exitCode=0 Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.883708 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzl2l" event={"ID":"31a015ae-51a2-488b-8bfd-4829857d6c7d","Type":"ContainerDied","Data":"9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba"} Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.894205 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.903049 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.924755 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.925152 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.935128 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.950314 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7vk2f"] Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.952071 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.958968 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.960478 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:05 crc kubenswrapper[4735]: E0215 20:19:05.961503 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.461473153 +0000 UTC m=+154.327488776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:05 crc kubenswrapper[4735]: I0215 20:19:05.965583 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7vk2f"] Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.053773 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.067377 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:06 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:06 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:06 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.067447 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.069418 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.069572 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.069797 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdrr6\" (UniqueName: \"kubernetes.io/projected/309558c5-0acb-4b54-aedf-bde3c482dec9-kube-api-access-vdrr6\") pod \"redhat-operators-7vk2f\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.070228 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-catalog-content\") pod \"redhat-operators-7vk2f\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.070371 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.070571 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-utilities\") pod \"redhat-operators-7vk2f\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.083320 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.583301598 +0000 UTC m=+154.449317221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.150858 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.174748 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.175109 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.175145 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.175180 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdrr6\" (UniqueName: \"kubernetes.io/projected/309558c5-0acb-4b54-aedf-bde3c482dec9-kube-api-access-vdrr6\") pod \"redhat-operators-7vk2f\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.175261 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-catalog-content\") pod \"redhat-operators-7vk2f\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.175315 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-utilities\") pod \"redhat-operators-7vk2f\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.176093 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-utilities\") pod \"redhat-operators-7vk2f\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.176606 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.676576051 +0000 UTC m=+154.542591864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.176888 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-catalog-content\") pod \"redhat-operators-7vk2f\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.177155 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.193788 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hm2zs"] Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.212385 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdrr6\" (UniqueName: \"kubernetes.io/projected/309558c5-0acb-4b54-aedf-bde3c482dec9-kube-api-access-vdrr6\") pod \"redhat-operators-7vk2f\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.245487 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.260833 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.278816 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.279270 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.779256775 +0000 UTC m=+154.645272398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.304013 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zwjpk"] Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.305237 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.313162 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.324866 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zwjpk"] Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.369538 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bxj7"] Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.383168 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.383603 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.883582506 +0000 UTC m=+154.749598119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.489218 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-catalog-content\") pod \"redhat-operators-zwjpk\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.489307 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhbg8\" (UniqueName: \"kubernetes.io/projected/aee39f47-6719-4de1-ab1c-5a8814345098-kube-api-access-xhbg8\") pod \"redhat-operators-zwjpk\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.489349 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.489379 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-utilities\") pod \"redhat-operators-zwjpk\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.489889 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:06.989872244 +0000 UTC m=+154.855887867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.591776 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.592160 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:07.092107204 +0000 UTC m=+154.958122827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.592681 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-catalog-content\") pod \"redhat-operators-zwjpk\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.592733 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhbg8\" (UniqueName: \"kubernetes.io/projected/aee39f47-6719-4de1-ab1c-5a8814345098-kube-api-access-xhbg8\") pod \"redhat-operators-zwjpk\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.592757 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.592777 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-utilities\") pod \"redhat-operators-zwjpk\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.593329 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-utilities\") pod \"redhat-operators-zwjpk\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.593719 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:07.093702091 +0000 UTC m=+154.959717714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.593779 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-catalog-content\") pod \"redhat-operators-zwjpk\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.617519 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhbg8\" (UniqueName: \"kubernetes.io/projected/aee39f47-6719-4de1-ab1c-5a8814345098-kube-api-access-xhbg8\") pod \"redhat-operators-zwjpk\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.658096 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.693915 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.694309 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:07.194290124 +0000 UTC m=+155.060305747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.796194 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.796646 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:07.296626388 +0000 UTC m=+155.162642011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.898974 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:06 crc kubenswrapper[4735]: E0215 20:19:06.899638 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:07.399616671 +0000 UTC m=+155.265632294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.901151 4735 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.956762 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bxj7" event={"ID":"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9","Type":"ContainerStarted","Data":"6338f5b6486ffef7db7f72ed1d5988f15f7252bc76fd8b8ccf2bac4a2210099f"} Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.959718 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm2zs" event={"ID":"59b9213c-60ff-4ecd-99d9-4c1fe670592c","Type":"ContainerStarted","Data":"e1dc20bdcf999f7f3033724a0132e6c49704fa9dc3dbf2a881e2bb1186fc2a8e"} Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.972902 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bb454db2-cace-46fe-8a7a-8ead6b1529f9","Type":"ContainerStarted","Data":"0444efb24014fe3a00c7c38709d7d00007c1521cab64ab392f2fcf42a056b285"} Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.975822 4735 generic.go:334] "Generic (PLEG): container finished" podID="4f5c12d8-7152-4299-bcad-c59201575285" containerID="3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c" exitCode=0 Feb 15 20:19:06 crc kubenswrapper[4735]: I0215 20:19:06.976533 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9ft4" event={"ID":"4f5c12d8-7152-4299-bcad-c59201575285","Type":"ContainerDied","Data":"3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c"} Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.000864 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:07 crc kubenswrapper[4735]: E0215 20:19:07.001530 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:07.501513572 +0000 UTC m=+155.367529185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.075211 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:07 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:07 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:07 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.075311 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.091002 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7vk2f"] Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.104628 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:07 crc kubenswrapper[4735]: E0215 20:19:07.104683 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-15 20:19:07.604661949 +0000 UTC m=+155.470677572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.105887 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:07 crc kubenswrapper[4735]: E0215 20:19:07.107638 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-15 20:19:07.607613674 +0000 UTC m=+155.473629297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9h4lw" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.143652 4735 patch_prober.go:28] interesting pod/apiserver-76f77b778f-jbnv7 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]log ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]etcd ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/generic-apiserver-start-informers ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/max-in-flight-filter ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 15 20:19:07 crc kubenswrapper[4735]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 15 20:19:07 crc kubenswrapper[4735]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/project.openshift.io-projectcache ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/openshift.io-startinformers ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 15 20:19:07 crc kubenswrapper[4735]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 15 20:19:07 crc kubenswrapper[4735]: livez check failed Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.143716 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" podUID="d3be498a-3362-4c71-85ac-e53b4889eec5" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.144050 4735 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-15T20:19:06.901188556Z","Handler":null,"Name":""} Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.163357 4735 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.163499 4735 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.216602 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.367744 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.424915 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.501565 4735 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.501616 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.582916 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.655219 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9h4lw\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.702421 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zwjpk"] Feb 15 20:19:07 crc kubenswrapper[4735]: W0215 20:19:07.769583 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaee39f47_6719_4de1_ab1c_5a8814345098.slice/crio-219a314d39da1d7be743247d50469ff73e5af97d6bae57a1b0e09cdabc0d3474 WatchSource:0}: Error finding container 219a314d39da1d7be743247d50469ff73e5af97d6bae57a1b0e09cdabc0d3474: Status 404 returned error can't find the container with id 219a314d39da1d7be743247d50469ff73e5af97d6bae57a1b0e09cdabc0d3474 Feb 15 20:19:07 crc kubenswrapper[4735]: I0215 20:19:07.953583 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.006676 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"aec9d74e-47ac-41a5-8a70-dca86963b5a9","Type":"ContainerStarted","Data":"24bad2f5d75725f7e75481a08145776dfbe8f97a5782c6339e98ef26b060e39b"} Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.010404 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zwjpk" event={"ID":"aee39f47-6719-4de1-ab1c-5a8814345098","Type":"ContainerStarted","Data":"219a314d39da1d7be743247d50469ff73e5af97d6bae57a1b0e09cdabc0d3474"} Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.066332 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:08 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:08 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:08 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.066870 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.079808 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" event={"ID":"fc8fe06e-5292-4b4c-9b20-b2a3f4fce798","Type":"ContainerStarted","Data":"fed2c4aed8f91e0c035202426f6f1d2f33ebcdc298ca12bc8de8d2e2fc7e7afd"} Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.097659 4735 generic.go:334] "Generic (PLEG): container finished" podID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerID="ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336" exitCode=0 Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.097785 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bxj7" event={"ID":"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9","Type":"ContainerDied","Data":"ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336"} Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.108967 4735 generic.go:334] "Generic (PLEG): container finished" podID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerID="613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26" exitCode=0 Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.109035 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vk2f" event={"ID":"309558c5-0acb-4b54-aedf-bde3c482dec9","Type":"ContainerDied","Data":"613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26"} Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.109063 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vk2f" event={"ID":"309558c5-0acb-4b54-aedf-bde3c482dec9","Type":"ContainerStarted","Data":"5dc97d1d176104400775f5bbfb7f9e3cda6b2980f294bfa976576c195737258a"} Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.116761 4735 generic.go:334] "Generic (PLEG): container finished" podID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerID="32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c" exitCode=0 Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.116802 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm2zs" event={"ID":"59b9213c-60ff-4ecd-99d9-4c1fe670592c","Type":"ContainerDied","Data":"32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c"} Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.121598 4735 generic.go:334] "Generic (PLEG): container finished" podID="bb454db2-cace-46fe-8a7a-8ead6b1529f9" containerID="d0528d2013a182d37244a030956787d643cbb37209dcf0919caf214fdbe4b135" exitCode=0 Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.121624 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bb454db2-cace-46fe-8a7a-8ead6b1529f9","Type":"ContainerDied","Data":"d0528d2013a182d37244a030956787d643cbb37209dcf0919caf214fdbe4b135"} Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.148162 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" podStartSLOduration=16.148134047 podStartE2EDuration="16.148134047s" podCreationTimestamp="2026-02-15 20:18:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:08.125409094 +0000 UTC m=+155.991424717" watchObservedRunningTime="2026-02-15 20:19:08.148134047 +0000 UTC m=+156.014149670" Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.810586 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9h4lw"] Feb 15 20:19:08 crc kubenswrapper[4735]: I0215 20:19:08.899007 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.092397 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:09 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:09 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:09 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.092469 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.176879 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"aec9d74e-47ac-41a5-8a70-dca86963b5a9","Type":"ContainerStarted","Data":"2b314c51b5f41dc3f3a82261b5611cac5f552570f933b7a966335fc200ed2a0d"} Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.181803 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" event={"ID":"d4b1b306-2224-4908-b9f6-484d76a77ee1","Type":"ContainerStarted","Data":"0b832c4933cf53e1ea10ce401b97a61c2a784e7be0b9621036f68dcaa5fafd7d"} Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.187480 4735 generic.go:334] "Generic (PLEG): container finished" podID="aee39f47-6719-4de1-ab1c-5a8814345098" containerID="1340a1595134c3cdfb3920a15102720a3ab52af2f6d4f1e33152861e659d0dd1" exitCode=0 Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.188030 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zwjpk" event={"ID":"aee39f47-6719-4de1-ab1c-5a8814345098","Type":"ContainerDied","Data":"1340a1595134c3cdfb3920a15102720a3ab52af2f6d4f1e33152861e659d0dd1"} Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.202210 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.202184149 podStartE2EDuration="4.202184149s" podCreationTimestamp="2026-02-15 20:19:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:09.201817569 +0000 UTC m=+157.067833192" watchObservedRunningTime="2026-02-15 20:19:09.202184149 +0000 UTC m=+157.068199772" Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.824903 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.918698 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kubelet-dir\") pod \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\" (UID: \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\") " Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.918814 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bb454db2-cace-46fe-8a7a-8ead6b1529f9" (UID: "bb454db2-cace-46fe-8a7a-8ead6b1529f9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.918879 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kube-api-access\") pod \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\" (UID: \"bb454db2-cace-46fe-8a7a-8ead6b1529f9\") " Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.919281 4735 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:19:09 crc kubenswrapper[4735]: I0215 20:19:09.928252 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bb454db2-cace-46fe-8a7a-8ead6b1529f9" (UID: "bb454db2-cace-46fe-8a7a-8ead6b1529f9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.031013 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb454db2-cace-46fe-8a7a-8ead6b1529f9-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.066279 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:10 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:10 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:10 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.066364 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.288990 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bb454db2-cace-46fe-8a7a-8ead6b1529f9","Type":"ContainerDied","Data":"0444efb24014fe3a00c7c38709d7d00007c1521cab64ab392f2fcf42a056b285"} Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.289054 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0444efb24014fe3a00c7c38709d7d00007c1521cab64ab392f2fcf42a056b285" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.289130 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.319304 4735 generic.go:334] "Generic (PLEG): container finished" podID="aec9d74e-47ac-41a5-8a70-dca86963b5a9" containerID="2b314c51b5f41dc3f3a82261b5611cac5f552570f933b7a966335fc200ed2a0d" exitCode=0 Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.319387 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"aec9d74e-47ac-41a5-8a70-dca86963b5a9","Type":"ContainerDied","Data":"2b314c51b5f41dc3f3a82261b5611cac5f552570f933b7a966335fc200ed2a0d"} Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.345304 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.356888 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" event={"ID":"d4b1b306-2224-4908-b9f6-484d76a77ee1","Type":"ContainerStarted","Data":"3388046baa5114834e8c40f562b69fc131769c6832fa605d306b2911327e3e7a"} Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.357632 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.364285 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fg9hk" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.393999 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" podStartSLOduration=136.393971033 podStartE2EDuration="2m16.393971033s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:10.393292434 +0000 UTC m=+158.259308077" watchObservedRunningTime="2026-02-15 20:19:10.393971033 +0000 UTC m=+158.259986656" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.628702 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:19:10 crc kubenswrapper[4735]: I0215 20:19:10.637508 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-jbnv7" Feb 15 20:19:11 crc kubenswrapper[4735]: I0215 20:19:11.056939 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:11 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:11 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:11 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:11 crc kubenswrapper[4735]: I0215 20:19:11.057541 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:11 crc kubenswrapper[4735]: I0215 20:19:11.231429 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-l8sf2" Feb 15 20:19:11 crc kubenswrapper[4735]: I0215 20:19:11.887959 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:11 crc kubenswrapper[4735]: I0215 20:19:11.980654 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kubelet-dir\") pod \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\" (UID: \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\") " Feb 15 20:19:11 crc kubenswrapper[4735]: I0215 20:19:11.980853 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kube-api-access\") pod \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\" (UID: \"aec9d74e-47ac-41a5-8a70-dca86963b5a9\") " Feb 15 20:19:11 crc kubenswrapper[4735]: I0215 20:19:11.982099 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "aec9d74e-47ac-41a5-8a70-dca86963b5a9" (UID: "aec9d74e-47ac-41a5-8a70-dca86963b5a9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:19:11 crc kubenswrapper[4735]: I0215 20:19:11.991236 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "aec9d74e-47ac-41a5-8a70-dca86963b5a9" (UID: "aec9d74e-47ac-41a5-8a70-dca86963b5a9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:19:12 crc kubenswrapper[4735]: I0215 20:19:12.061056 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:12 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:12 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:12 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:12 crc kubenswrapper[4735]: I0215 20:19:12.061142 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:12 crc kubenswrapper[4735]: I0215 20:19:12.082799 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 15 20:19:12 crc kubenswrapper[4735]: I0215 20:19:12.082838 4735 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aec9d74e-47ac-41a5-8a70-dca86963b5a9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:19:12 crc kubenswrapper[4735]: I0215 20:19:12.456654 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 15 20:19:12 crc kubenswrapper[4735]: I0215 20:19:12.456639 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"aec9d74e-47ac-41a5-8a70-dca86963b5a9","Type":"ContainerDied","Data":"24bad2f5d75725f7e75481a08145776dfbe8f97a5782c6339e98ef26b060e39b"} Feb 15 20:19:12 crc kubenswrapper[4735]: I0215 20:19:12.456987 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24bad2f5d75725f7e75481a08145776dfbe8f97a5782c6339e98ef26b060e39b" Feb 15 20:19:13 crc kubenswrapper[4735]: I0215 20:19:13.057292 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:13 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:13 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:13 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:13 crc kubenswrapper[4735]: I0215 20:19:13.057371 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.056033 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:14 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:14 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:14 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.056118 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.794589 4735 patch_prober.go:28] interesting pod/console-f9d7485db-k64kb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.794924 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-k64kb" podUID="e74e438d-61ef-403a-b958-8704f195f0ce" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.815244 4735 generic.go:334] "Generic (PLEG): container finished" podID="a930872e-2679-46c9-ac6f-22bffe1b8d66" containerID="fd2feb5ae65c846e3bfc78616e103db5f08dd6d485bc8e98f0fea784efdca889" exitCode=0 Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.815302 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" event={"ID":"a930872e-2679-46c9-ac6f-22bffe1b8d66","Type":"ContainerDied","Data":"fd2feb5ae65c846e3bfc78616e103db5f08dd6d485bc8e98f0fea784efdca889"} Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.957391 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-zjpgj container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.957393 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-zjpgj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.957478 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-zjpgj" podUID="0faea847-63f9-488b-9f3a-080e36fd9025" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 15 20:19:14 crc kubenswrapper[4735]: I0215 20:19:14.957545 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zjpgj" podUID="0faea847-63f9-488b-9f3a-080e36fd9025" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 15 20:19:15 crc kubenswrapper[4735]: I0215 20:19:15.055347 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:15 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:15 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:15 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:15 crc kubenswrapper[4735]: I0215 20:19:15.055420 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.056736 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 15 20:19:16 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Feb 15 20:19:16 crc kubenswrapper[4735]: [+]process-running ok Feb 15 20:19:16 crc kubenswrapper[4735]: healthz check failed Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.057336 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.668041 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.672512 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a930872e-2679-46c9-ac6f-22bffe1b8d66-config-volume\") pod \"a930872e-2679-46c9-ac6f-22bffe1b8d66\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.672672 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjm5b\" (UniqueName: \"kubernetes.io/projected/a930872e-2679-46c9-ac6f-22bffe1b8d66-kube-api-access-hjm5b\") pod \"a930872e-2679-46c9-ac6f-22bffe1b8d66\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.672735 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a930872e-2679-46c9-ac6f-22bffe1b8d66-secret-volume\") pod \"a930872e-2679-46c9-ac6f-22bffe1b8d66\" (UID: \"a930872e-2679-46c9-ac6f-22bffe1b8d66\") " Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.677841 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a930872e-2679-46c9-ac6f-22bffe1b8d66-config-volume" (OuterVolumeSpecName: "config-volume") pod "a930872e-2679-46c9-ac6f-22bffe1b8d66" (UID: "a930872e-2679-46c9-ac6f-22bffe1b8d66"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.703649 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a930872e-2679-46c9-ac6f-22bffe1b8d66-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a930872e-2679-46c9-ac6f-22bffe1b8d66" (UID: "a930872e-2679-46c9-ac6f-22bffe1b8d66"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.706846 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a930872e-2679-46c9-ac6f-22bffe1b8d66-kube-api-access-hjm5b" (OuterVolumeSpecName: "kube-api-access-hjm5b") pod "a930872e-2679-46c9-ac6f-22bffe1b8d66" (UID: "a930872e-2679-46c9-ac6f-22bffe1b8d66"). InnerVolumeSpecName "kube-api-access-hjm5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.775894 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a930872e-2679-46c9-ac6f-22bffe1b8d66-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.775978 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a930872e-2679-46c9-ac6f-22bffe1b8d66-config-volume\") on node \"crc\" DevicePath \"\"" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.775992 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjm5b\" (UniqueName: \"kubernetes.io/projected/a930872e-2679-46c9-ac6f-22bffe1b8d66-kube-api-access-hjm5b\") on node \"crc\" DevicePath \"\"" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.872896 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.891318 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h" event={"ID":"a930872e-2679-46c9-ac6f-22bffe1b8d66","Type":"ContainerDied","Data":"262a806652391feb5cb48d58492781291011665125e16c3339dca4a80856a9bc"} Feb 15 20:19:16 crc kubenswrapper[4735]: I0215 20:19:16.891410 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="262a806652391feb5cb48d58492781291011665125e16c3339dca4a80856a9bc" Feb 15 20:19:17 crc kubenswrapper[4735]: I0215 20:19:17.056131 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:19:17 crc kubenswrapper[4735]: I0215 20:19:17.060784 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lj984" Feb 15 20:19:17 crc kubenswrapper[4735]: I0215 20:19:17.604254 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:19:17 crc kubenswrapper[4735]: I0215 20:19:17.626380 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96681837-f2a3-42d2-9653-a3997f396291-metrics-certs\") pod \"network-metrics-daemon-7vfs4\" (UID: \"96681837-f2a3-42d2-9653-a3997f396291\") " pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:19:17 crc kubenswrapper[4735]: I0215 20:19:17.830887 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7vfs4" Feb 15 20:19:18 crc kubenswrapper[4735]: I0215 20:19:18.408887 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7vfs4"] Feb 15 20:19:18 crc kubenswrapper[4735]: W0215 20:19:18.498505 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96681837_f2a3_42d2_9653_a3997f396291.slice/crio-4b3065a20ff032b6964cfacb7d1513bdd297ba391ea4763ed4acdf645ba014dc WatchSource:0}: Error finding container 4b3065a20ff032b6964cfacb7d1513bdd297ba391ea4763ed4acdf645ba014dc: Status 404 returned error can't find the container with id 4b3065a20ff032b6964cfacb7d1513bdd297ba391ea4763ed4acdf645ba014dc Feb 15 20:19:18 crc kubenswrapper[4735]: I0215 20:19:18.975667 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" event={"ID":"96681837-f2a3-42d2-9653-a3997f396291","Type":"ContainerStarted","Data":"4b3065a20ff032b6964cfacb7d1513bdd297ba391ea4763ed4acdf645ba014dc"} Feb 15 20:19:19 crc kubenswrapper[4735]: I0215 20:19:19.680319 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:19:19 crc kubenswrapper[4735]: I0215 20:19:19.681783 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:19:20 crc kubenswrapper[4735]: I0215 20:19:20.001421 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" event={"ID":"96681837-f2a3-42d2-9653-a3997f396291","Type":"ContainerStarted","Data":"fd51a07d92be146ed18530d98101dfa3cda03e006340220e1db4921c865218eb"} Feb 15 20:19:24 crc kubenswrapper[4735]: I0215 20:19:24.797071 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:19:24 crc kubenswrapper[4735]: I0215 20:19:24.802923 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:19:24 crc kubenswrapper[4735]: I0215 20:19:24.962631 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-zjpgj" Feb 15 20:19:27 crc kubenswrapper[4735]: I0215 20:19:27.965580 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:19:35 crc kubenswrapper[4735]: I0215 20:19:35.584784 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-v7mzq" Feb 15 20:19:41 crc kubenswrapper[4735]: I0215 20:19:41.358634 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.823029 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 15 20:19:42 crc kubenswrapper[4735]: E0215 20:19:42.823891 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb454db2-cace-46fe-8a7a-8ead6b1529f9" containerName="pruner" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.823903 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb454db2-cace-46fe-8a7a-8ead6b1529f9" containerName="pruner" Feb 15 20:19:42 crc kubenswrapper[4735]: E0215 20:19:42.823914 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a930872e-2679-46c9-ac6f-22bffe1b8d66" containerName="collect-profiles" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.823922 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a930872e-2679-46c9-ac6f-22bffe1b8d66" containerName="collect-profiles" Feb 15 20:19:42 crc kubenswrapper[4735]: E0215 20:19:42.823929 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec9d74e-47ac-41a5-8a70-dca86963b5a9" containerName="pruner" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.823935 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec9d74e-47ac-41a5-8a70-dca86963b5a9" containerName="pruner" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.824083 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a930872e-2679-46c9-ac6f-22bffe1b8d66" containerName="collect-profiles" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.824096 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="aec9d74e-47ac-41a5-8a70-dca86963b5a9" containerName="pruner" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.824106 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb454db2-cace-46fe-8a7a-8ead6b1529f9" containerName="pruner" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.824920 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.831558 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.831782 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.839736 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.973765 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d90f00-7bbc-4358-9958-bf32d1bb967a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"30d90f00-7bbc-4358-9958-bf32d1bb967a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:42 crc kubenswrapper[4735]: I0215 20:19:42.973821 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d90f00-7bbc-4358-9958-bf32d1bb967a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"30d90f00-7bbc-4358-9958-bf32d1bb967a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:43 crc kubenswrapper[4735]: I0215 20:19:43.074966 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d90f00-7bbc-4358-9958-bf32d1bb967a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"30d90f00-7bbc-4358-9958-bf32d1bb967a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:43 crc kubenswrapper[4735]: I0215 20:19:43.075006 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d90f00-7bbc-4358-9958-bf32d1bb967a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"30d90f00-7bbc-4358-9958-bf32d1bb967a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:43 crc kubenswrapper[4735]: I0215 20:19:43.075187 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d90f00-7bbc-4358-9958-bf32d1bb967a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"30d90f00-7bbc-4358-9958-bf32d1bb967a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:43 crc kubenswrapper[4735]: I0215 20:19:43.095936 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d90f00-7bbc-4358-9958-bf32d1bb967a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"30d90f00-7bbc-4358-9958-bf32d1bb967a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:43 crc kubenswrapper[4735]: I0215 20:19:43.161144 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:46 crc kubenswrapper[4735]: I0215 20:19:46.827086 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 15 20:19:46 crc kubenswrapper[4735]: I0215 20:19:46.829569 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:46 crc kubenswrapper[4735]: I0215 20:19:46.832634 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 15 20:19:46 crc kubenswrapper[4735]: I0215 20:19:46.925922 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa80450c-ee41-4f6f-813b-ba7556c751c0-kube-api-access\") pod \"installer-9-crc\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:46 crc kubenswrapper[4735]: I0215 20:19:46.926076 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-var-lock\") pod \"installer-9-crc\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:46 crc kubenswrapper[4735]: I0215 20:19:46.926206 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-kubelet-dir\") pod \"installer-9-crc\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:47 crc kubenswrapper[4735]: I0215 20:19:47.027068 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa80450c-ee41-4f6f-813b-ba7556c751c0-kube-api-access\") pod \"installer-9-crc\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:47 crc kubenswrapper[4735]: I0215 20:19:47.027725 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-var-lock\") pod \"installer-9-crc\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:47 crc kubenswrapper[4735]: I0215 20:19:47.027767 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-kubelet-dir\") pod \"installer-9-crc\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:47 crc kubenswrapper[4735]: I0215 20:19:47.027840 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-kubelet-dir\") pod \"installer-9-crc\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:47 crc kubenswrapper[4735]: I0215 20:19:47.027847 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-var-lock\") pod \"installer-9-crc\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:47 crc kubenswrapper[4735]: I0215 20:19:47.062317 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa80450c-ee41-4f6f-813b-ba7556c751c0-kube-api-access\") pod \"installer-9-crc\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:47 crc kubenswrapper[4735]: I0215 20:19:47.159569 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:19:49 crc kubenswrapper[4735]: I0215 20:19:49.679360 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:19:49 crc kubenswrapper[4735]: I0215 20:19:49.679812 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.362240 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.362474 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55xzs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4bxj7_openshift-marketplace(853b0875-62d2-42fd-ad1a-c8fe3d16e8b9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.364087 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4bxj7" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.369721 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.370176 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wsvrk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lzl2l_openshift-marketplace(31a015ae-51a2-488b-8bfd-4829857d6c7d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.371373 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lzl2l" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.397426 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.397593 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vdrr6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-7vk2f_openshift-marketplace(309558c5-0acb-4b54-aedf-bde3c482dec9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.399684 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-7vk2f" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.556186 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.556638 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xhbg8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zwjpk_openshift-marketplace(aee39f47-6719-4de1-ab1c-5a8814345098): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 15 20:19:50 crc kubenswrapper[4735]: E0215 20:19:50.557837 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zwjpk" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" Feb 15 20:19:50 crc kubenswrapper[4735]: I0215 20:19:50.798923 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 15 20:19:50 crc kubenswrapper[4735]: W0215 20:19:50.804207 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod30d90f00_7bbc_4358_9958_bf32d1bb967a.slice/crio-81796cc8e768f747a64aca2ab1f010c31a74dd31d6d9e8d55c6f701905fe38fc WatchSource:0}: Error finding container 81796cc8e768f747a64aca2ab1f010c31a74dd31d6d9e8d55c6f701905fe38fc: Status 404 returned error can't find the container with id 81796cc8e768f747a64aca2ab1f010c31a74dd31d6d9e8d55c6f701905fe38fc Feb 15 20:19:50 crc kubenswrapper[4735]: I0215 20:19:50.918841 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.237378 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fa80450c-ee41-4f6f-813b-ba7556c751c0","Type":"ContainerStarted","Data":"a21272bfc8258468f2df11a28a6c17806b67fa0cf574d2c2782084eb6d6071f0"} Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.240356 4735 generic.go:334] "Generic (PLEG): container finished" podID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerID="bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600" exitCode=0 Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.240401 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm2zs" event={"ID":"59b9213c-60ff-4ecd-99d9-4c1fe670592c","Type":"ContainerDied","Data":"bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600"} Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.244331 4735 generic.go:334] "Generic (PLEG): container finished" podID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerID="8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28" exitCode=0 Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.244378 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ml7r" event={"ID":"bf815092-d21d-4b8b-bbfa-fd11550ffb26","Type":"ContainerDied","Data":"8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28"} Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.247069 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"30d90f00-7bbc-4358-9958-bf32d1bb967a","Type":"ContainerStarted","Data":"81796cc8e768f747a64aca2ab1f010c31a74dd31d6d9e8d55c6f701905fe38fc"} Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.250335 4735 generic.go:334] "Generic (PLEG): container finished" podID="4f5c12d8-7152-4299-bcad-c59201575285" containerID="77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b" exitCode=0 Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.250414 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9ft4" event={"ID":"4f5c12d8-7152-4299-bcad-c59201575285","Type":"ContainerDied","Data":"77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b"} Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.257471 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7vfs4" event={"ID":"96681837-f2a3-42d2-9653-a3997f396291","Type":"ContainerStarted","Data":"489afec704cab8d266f07d5cbb7ea74e53518c299e1d368d632583ea93d7c013"} Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.262477 4735 generic.go:334] "Generic (PLEG): container finished" podID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerID="47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594" exitCode=0 Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.264632 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2cm8" event={"ID":"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23","Type":"ContainerDied","Data":"47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594"} Feb 15 20:19:51 crc kubenswrapper[4735]: E0215 20:19:51.270404 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zwjpk" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" Feb 15 20:19:51 crc kubenswrapper[4735]: E0215 20:19:51.273139 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-7vk2f" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" Feb 15 20:19:51 crc kubenswrapper[4735]: E0215 20:19:51.284246 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lzl2l" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" Feb 15 20:19:51 crc kubenswrapper[4735]: I0215 20:19:51.285514 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7vfs4" podStartSLOduration=177.285497685 podStartE2EDuration="2m57.285497685s" podCreationTimestamp="2026-02-15 20:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:51.281975866 +0000 UTC m=+199.147991489" watchObservedRunningTime="2026-02-15 20:19:51.285497685 +0000 UTC m=+199.151513308" Feb 15 20:19:51 crc kubenswrapper[4735]: E0215 20:19:51.291757 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4bxj7" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" Feb 15 20:19:52 crc kubenswrapper[4735]: I0215 20:19:52.268662 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fa80450c-ee41-4f6f-813b-ba7556c751c0","Type":"ContainerStarted","Data":"4d0111dfd93f2c773dcd658b295fdde009d76f0cf2af8bbfc89cb9f45a5448ec"} Feb 15 20:19:52 crc kubenswrapper[4735]: I0215 20:19:52.270142 4735 generic.go:334] "Generic (PLEG): container finished" podID="30d90f00-7bbc-4358-9958-bf32d1bb967a" containerID="58d7a39a10cabb3449a5f1d059f9151a26bf236d0d467246fa28bb1de89ab86d" exitCode=0 Feb 15 20:19:52 crc kubenswrapper[4735]: I0215 20:19:52.270174 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"30d90f00-7bbc-4358-9958-bf32d1bb967a","Type":"ContainerDied","Data":"58d7a39a10cabb3449a5f1d059f9151a26bf236d0d467246fa28bb1de89ab86d"} Feb 15 20:19:52 crc kubenswrapper[4735]: I0215 20:19:52.290386 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=6.290364952 podStartE2EDuration="6.290364952s" podCreationTimestamp="2026-02-15 20:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:19:52.285467154 +0000 UTC m=+200.151482777" watchObservedRunningTime="2026-02-15 20:19:52.290364952 +0000 UTC m=+200.156380575" Feb 15 20:19:53 crc kubenswrapper[4735]: I0215 20:19:53.280608 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ml7r" event={"ID":"bf815092-d21d-4b8b-bbfa-fd11550ffb26","Type":"ContainerStarted","Data":"0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23"} Feb 15 20:19:53 crc kubenswrapper[4735]: I0215 20:19:53.569368 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:53 crc kubenswrapper[4735]: I0215 20:19:53.585264 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d90f00-7bbc-4358-9958-bf32d1bb967a-kube-api-access\") pod \"30d90f00-7bbc-4358-9958-bf32d1bb967a\" (UID: \"30d90f00-7bbc-4358-9958-bf32d1bb967a\") " Feb 15 20:19:53 crc kubenswrapper[4735]: I0215 20:19:53.585319 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d90f00-7bbc-4358-9958-bf32d1bb967a-kubelet-dir\") pod \"30d90f00-7bbc-4358-9958-bf32d1bb967a\" (UID: \"30d90f00-7bbc-4358-9958-bf32d1bb967a\") " Feb 15 20:19:53 crc kubenswrapper[4735]: I0215 20:19:53.585467 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30d90f00-7bbc-4358-9958-bf32d1bb967a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "30d90f00-7bbc-4358-9958-bf32d1bb967a" (UID: "30d90f00-7bbc-4358-9958-bf32d1bb967a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:19:53 crc kubenswrapper[4735]: I0215 20:19:53.585946 4735 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d90f00-7bbc-4358-9958-bf32d1bb967a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:19:53 crc kubenswrapper[4735]: I0215 20:19:53.593401 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d90f00-7bbc-4358-9958-bf32d1bb967a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "30d90f00-7bbc-4358-9958-bf32d1bb967a" (UID: "30d90f00-7bbc-4358-9958-bf32d1bb967a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:19:53 crc kubenswrapper[4735]: I0215 20:19:53.686863 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d90f00-7bbc-4358-9958-bf32d1bb967a-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 15 20:19:53 crc kubenswrapper[4735]: I0215 20:19:53.946156 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r89v"] Feb 15 20:19:54 crc kubenswrapper[4735]: I0215 20:19:54.288496 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"30d90f00-7bbc-4358-9958-bf32d1bb967a","Type":"ContainerDied","Data":"81796cc8e768f747a64aca2ab1f010c31a74dd31d6d9e8d55c6f701905fe38fc"} Feb 15 20:19:54 crc kubenswrapper[4735]: I0215 20:19:54.288872 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81796cc8e768f747a64aca2ab1f010c31a74dd31d6d9e8d55c6f701905fe38fc" Feb 15 20:19:54 crc kubenswrapper[4735]: I0215 20:19:54.288549 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 15 20:19:54 crc kubenswrapper[4735]: I0215 20:19:54.318781 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7ml7r" podStartSLOduration=5.259496138 podStartE2EDuration="52.318764743s" podCreationTimestamp="2026-02-15 20:19:02 +0000 UTC" firstStartedPulling="2026-02-15 20:19:05.844097696 +0000 UTC m=+153.710113319" lastFinishedPulling="2026-02-15 20:19:52.903366291 +0000 UTC m=+200.769381924" observedRunningTime="2026-02-15 20:19:54.316434547 +0000 UTC m=+202.182450170" watchObservedRunningTime="2026-02-15 20:19:54.318764743 +0000 UTC m=+202.184780366" Feb 15 20:19:55 crc kubenswrapper[4735]: I0215 20:19:55.296281 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2cm8" event={"ID":"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23","Type":"ContainerStarted","Data":"e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99"} Feb 15 20:19:55 crc kubenswrapper[4735]: I0215 20:19:55.313914 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z2cm8" podStartSLOduration=5.02816066 podStartE2EDuration="53.313893367s" podCreationTimestamp="2026-02-15 20:19:02 +0000 UTC" firstStartedPulling="2026-02-15 20:19:05.787259812 +0000 UTC m=+153.653275435" lastFinishedPulling="2026-02-15 20:19:54.072992519 +0000 UTC m=+201.939008142" observedRunningTime="2026-02-15 20:19:55.312284552 +0000 UTC m=+203.178300175" watchObservedRunningTime="2026-02-15 20:19:55.313893367 +0000 UTC m=+203.179908990" Feb 15 20:19:56 crc kubenswrapper[4735]: I0215 20:19:56.303788 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm2zs" event={"ID":"59b9213c-60ff-4ecd-99d9-4c1fe670592c","Type":"ContainerStarted","Data":"8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b"} Feb 15 20:19:56 crc kubenswrapper[4735]: I0215 20:19:56.306659 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9ft4" event={"ID":"4f5c12d8-7152-4299-bcad-c59201575285","Type":"ContainerStarted","Data":"987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3"} Feb 15 20:19:56 crc kubenswrapper[4735]: I0215 20:19:56.321509 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hm2zs" podStartSLOduration=5.218724748 podStartE2EDuration="52.321492331s" podCreationTimestamp="2026-02-15 20:19:04 +0000 UTC" firstStartedPulling="2026-02-15 20:19:08.118167705 +0000 UTC m=+155.984183328" lastFinishedPulling="2026-02-15 20:19:55.220935288 +0000 UTC m=+203.086950911" observedRunningTime="2026-02-15 20:19:56.320753611 +0000 UTC m=+204.186769234" watchObservedRunningTime="2026-02-15 20:19:56.321492331 +0000 UTC m=+204.187507954" Feb 15 20:19:56 crc kubenswrapper[4735]: I0215 20:19:56.341000 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f9ft4" podStartSLOduration=5.151360753 podStartE2EDuration="53.340982561s" podCreationTimestamp="2026-02-15 20:19:03 +0000 UTC" firstStartedPulling="2026-02-15 20:19:06.98229864 +0000 UTC m=+154.848314263" lastFinishedPulling="2026-02-15 20:19:55.171920448 +0000 UTC m=+203.037936071" observedRunningTime="2026-02-15 20:19:56.340887448 +0000 UTC m=+204.206903071" watchObservedRunningTime="2026-02-15 20:19:56.340982561 +0000 UTC m=+204.206998184" Feb 15 20:20:02 crc kubenswrapper[4735]: I0215 20:20:02.956759 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:20:02 crc kubenswrapper[4735]: I0215 20:20:02.957353 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:20:03 crc kubenswrapper[4735]: I0215 20:20:03.327590 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:20:03 crc kubenswrapper[4735]: I0215 20:20:03.327976 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:20:03 crc kubenswrapper[4735]: I0215 20:20:03.487052 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:20:03 crc kubenswrapper[4735]: I0215 20:20:03.488626 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:20:03 crc kubenswrapper[4735]: I0215 20:20:03.534848 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:20:03 crc kubenswrapper[4735]: I0215 20:20:03.549015 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:20:03 crc kubenswrapper[4735]: I0215 20:20:03.820170 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:20:03 crc kubenswrapper[4735]: I0215 20:20:03.820233 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:20:03 crc kubenswrapper[4735]: I0215 20:20:03.875456 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:20:04 crc kubenswrapper[4735]: I0215 20:20:04.385940 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7ml7r"] Feb 15 20:20:04 crc kubenswrapper[4735]: I0215 20:20:04.400476 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:20:04 crc kubenswrapper[4735]: I0215 20:20:04.922869 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:20:04 crc kubenswrapper[4735]: I0215 20:20:04.922925 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:20:04 crc kubenswrapper[4735]: I0215 20:20:04.967098 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:20:05 crc kubenswrapper[4735]: I0215 20:20:05.356224 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7ml7r" podUID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerName="registry-server" containerID="cri-o://0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23" gracePeriod=2 Feb 15 20:20:05 crc kubenswrapper[4735]: I0215 20:20:05.388568 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.181644 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f9ft4"] Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.230189 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.344534 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-catalog-content\") pod \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.344597 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-utilities\") pod \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.344640 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnlcx\" (UniqueName: \"kubernetes.io/projected/bf815092-d21d-4b8b-bbfa-fd11550ffb26-kube-api-access-vnlcx\") pod \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\" (UID: \"bf815092-d21d-4b8b-bbfa-fd11550ffb26\") " Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.348123 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-utilities" (OuterVolumeSpecName: "utilities") pod "bf815092-d21d-4b8b-bbfa-fd11550ffb26" (UID: "bf815092-d21d-4b8b-bbfa-fd11550ffb26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.353062 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf815092-d21d-4b8b-bbfa-fd11550ffb26-kube-api-access-vnlcx" (OuterVolumeSpecName: "kube-api-access-vnlcx") pod "bf815092-d21d-4b8b-bbfa-fd11550ffb26" (UID: "bf815092-d21d-4b8b-bbfa-fd11550ffb26"). InnerVolumeSpecName "kube-api-access-vnlcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.374097 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vk2f" event={"ID":"309558c5-0acb-4b54-aedf-bde3c482dec9","Type":"ContainerStarted","Data":"b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50"} Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.376989 4735 generic.go:334] "Generic (PLEG): container finished" podID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerID="0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23" exitCode=0 Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.377036 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ml7r" event={"ID":"bf815092-d21d-4b8b-bbfa-fd11550ffb26","Type":"ContainerDied","Data":"0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23"} Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.377052 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ml7r" event={"ID":"bf815092-d21d-4b8b-bbfa-fd11550ffb26","Type":"ContainerDied","Data":"d26a1ac054ce4dd551726ca95029c8a1899acc5e4ffaa4c806e4caa0aba77e4b"} Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.377068 4735 scope.go:117] "RemoveContainer" containerID="0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.377155 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ml7r" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.381179 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f9ft4" podUID="4f5c12d8-7152-4299-bcad-c59201575285" containerName="registry-server" containerID="cri-o://987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3" gracePeriod=2 Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.381260 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzl2l" event={"ID":"31a015ae-51a2-488b-8bfd-4829857d6c7d","Type":"ContainerStarted","Data":"ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9"} Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.402969 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf815092-d21d-4b8b-bbfa-fd11550ffb26" (UID: "bf815092-d21d-4b8b-bbfa-fd11550ffb26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.423271 4735 scope.go:117] "RemoveContainer" containerID="8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.447745 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.447775 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf815092-d21d-4b8b-bbfa-fd11550ffb26-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.447785 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnlcx\" (UniqueName: \"kubernetes.io/projected/bf815092-d21d-4b8b-bbfa-fd11550ffb26-kube-api-access-vnlcx\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.487970 4735 scope.go:117] "RemoveContainer" containerID="1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.505105 4735 scope.go:117] "RemoveContainer" containerID="0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23" Feb 15 20:20:06 crc kubenswrapper[4735]: E0215 20:20:06.505572 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23\": container with ID starting with 0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23 not found: ID does not exist" containerID="0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.505621 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23"} err="failed to get container status \"0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23\": rpc error: code = NotFound desc = could not find container \"0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23\": container with ID starting with 0e42b26d599f09895fa14de45fff3460536d61ae277a20c8c92b36d90347ff23 not found: ID does not exist" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.505663 4735 scope.go:117] "RemoveContainer" containerID="8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28" Feb 15 20:20:06 crc kubenswrapper[4735]: E0215 20:20:06.506401 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28\": container with ID starting with 8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28 not found: ID does not exist" containerID="8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.506458 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28"} err="failed to get container status \"8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28\": rpc error: code = NotFound desc = could not find container \"8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28\": container with ID starting with 8dfac9de3451eafdaa97fe93ab9e6001eef6f378ef3103dfc2b9d9e885f43a28 not found: ID does not exist" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.506499 4735 scope.go:117] "RemoveContainer" containerID="1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8" Feb 15 20:20:06 crc kubenswrapper[4735]: E0215 20:20:06.507491 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8\": container with ID starting with 1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8 not found: ID does not exist" containerID="1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.507522 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8"} err="failed to get container status \"1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8\": rpc error: code = NotFound desc = could not find container \"1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8\": container with ID starting with 1d5210754354c9d6c8cb5fa4b51e6a8172e89d672cb99814bf970cd602b9efe8 not found: ID does not exist" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.709759 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7ml7r"] Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.722375 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7ml7r"] Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.766031 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.855093 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-catalog-content\") pod \"4f5c12d8-7152-4299-bcad-c59201575285\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.855212 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27lfn\" (UniqueName: \"kubernetes.io/projected/4f5c12d8-7152-4299-bcad-c59201575285-kube-api-access-27lfn\") pod \"4f5c12d8-7152-4299-bcad-c59201575285\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.855265 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-utilities\") pod \"4f5c12d8-7152-4299-bcad-c59201575285\" (UID: \"4f5c12d8-7152-4299-bcad-c59201575285\") " Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.856066 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-utilities" (OuterVolumeSpecName: "utilities") pod "4f5c12d8-7152-4299-bcad-c59201575285" (UID: "4f5c12d8-7152-4299-bcad-c59201575285"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.860500 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f5c12d8-7152-4299-bcad-c59201575285-kube-api-access-27lfn" (OuterVolumeSpecName: "kube-api-access-27lfn") pod "4f5c12d8-7152-4299-bcad-c59201575285" (UID: "4f5c12d8-7152-4299-bcad-c59201575285"). InnerVolumeSpecName "kube-api-access-27lfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.892529 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" path="/var/lib/kubelet/pods/bf815092-d21d-4b8b-bbfa-fd11550ffb26/volumes" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.907029 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f5c12d8-7152-4299-bcad-c59201575285" (UID: "4f5c12d8-7152-4299-bcad-c59201575285"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.956249 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.956294 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27lfn\" (UniqueName: \"kubernetes.io/projected/4f5c12d8-7152-4299-bcad-c59201575285-kube-api-access-27lfn\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:06 crc kubenswrapper[4735]: I0215 20:20:06.956308 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5c12d8-7152-4299-bcad-c59201575285-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.389239 4735 generic.go:334] "Generic (PLEG): container finished" podID="4f5c12d8-7152-4299-bcad-c59201575285" containerID="987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3" exitCode=0 Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.389311 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9ft4" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.389313 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9ft4" event={"ID":"4f5c12d8-7152-4299-bcad-c59201575285","Type":"ContainerDied","Data":"987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3"} Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.389370 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9ft4" event={"ID":"4f5c12d8-7152-4299-bcad-c59201575285","Type":"ContainerDied","Data":"2aa4b66863df797b15a8d564c7a7fbd345ca09ef4bda4c56a0490a4c008948f4"} Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.389396 4735 scope.go:117] "RemoveContainer" containerID="987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.391736 4735 generic.go:334] "Generic (PLEG): container finished" podID="aee39f47-6719-4de1-ab1c-5a8814345098" containerID="5eef4918730b513010d0ee1f4e80e8f651f544920ad9294d46abc031ff0f3c1d" exitCode=0 Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.391805 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zwjpk" event={"ID":"aee39f47-6719-4de1-ab1c-5a8814345098","Type":"ContainerDied","Data":"5eef4918730b513010d0ee1f4e80e8f651f544920ad9294d46abc031ff0f3c1d"} Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.395304 4735 generic.go:334] "Generic (PLEG): container finished" podID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerID="ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9" exitCode=0 Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.395396 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzl2l" event={"ID":"31a015ae-51a2-488b-8bfd-4829857d6c7d","Type":"ContainerDied","Data":"ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9"} Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.404125 4735 scope.go:117] "RemoveContainer" containerID="77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.406330 4735 generic.go:334] "Generic (PLEG): container finished" podID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerID="68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300" exitCode=0 Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.406396 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bxj7" event={"ID":"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9","Type":"ContainerDied","Data":"68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300"} Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.422938 4735 generic.go:334] "Generic (PLEG): container finished" podID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerID="b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50" exitCode=0 Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.423513 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vk2f" event={"ID":"309558c5-0acb-4b54-aedf-bde3c482dec9","Type":"ContainerDied","Data":"b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50"} Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.457388 4735 scope.go:117] "RemoveContainer" containerID="3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.479459 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f9ft4"] Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.490056 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f9ft4"] Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.509139 4735 scope.go:117] "RemoveContainer" containerID="987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3" Feb 15 20:20:07 crc kubenswrapper[4735]: E0215 20:20:07.509838 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3\": container with ID starting with 987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3 not found: ID does not exist" containerID="987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.509880 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3"} err="failed to get container status \"987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3\": rpc error: code = NotFound desc = could not find container \"987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3\": container with ID starting with 987f54f6b18660d4da7a2ad06da2d45a86abf161f3abc73151d42825664944f3 not found: ID does not exist" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.509908 4735 scope.go:117] "RemoveContainer" containerID="77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b" Feb 15 20:20:07 crc kubenswrapper[4735]: E0215 20:20:07.510300 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b\": container with ID starting with 77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b not found: ID does not exist" containerID="77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.510537 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b"} err="failed to get container status \"77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b\": rpc error: code = NotFound desc = could not find container \"77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b\": container with ID starting with 77c691be2320ab5befc30e4db824e8856a76c0efe0a6811deac33185dc55175b not found: ID does not exist" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.510714 4735 scope.go:117] "RemoveContainer" containerID="3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c" Feb 15 20:20:07 crc kubenswrapper[4735]: E0215 20:20:07.511157 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c\": container with ID starting with 3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c not found: ID does not exist" containerID="3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c" Feb 15 20:20:07 crc kubenswrapper[4735]: I0215 20:20:07.511205 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c"} err="failed to get container status \"3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c\": rpc error: code = NotFound desc = could not find container \"3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c\": container with ID starting with 3ac1ae0bb23b62110e26819cb0839f5c8c8e6ec831c9086b023d381737699e3c not found: ID does not exist" Feb 15 20:20:08 crc kubenswrapper[4735]: I0215 20:20:08.435958 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vk2f" event={"ID":"309558c5-0acb-4b54-aedf-bde3c482dec9","Type":"ContainerStarted","Data":"1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82"} Feb 15 20:20:08 crc kubenswrapper[4735]: I0215 20:20:08.439405 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zwjpk" event={"ID":"aee39f47-6719-4de1-ab1c-5a8814345098","Type":"ContainerStarted","Data":"21973c3c069f2bb78afe17f168d610d1c6eb8f69e3c64949a1366455012076b2"} Feb 15 20:20:08 crc kubenswrapper[4735]: I0215 20:20:08.444602 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzl2l" event={"ID":"31a015ae-51a2-488b-8bfd-4829857d6c7d","Type":"ContainerStarted","Data":"ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112"} Feb 15 20:20:08 crc kubenswrapper[4735]: I0215 20:20:08.446781 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bxj7" event={"ID":"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9","Type":"ContainerStarted","Data":"4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77"} Feb 15 20:20:08 crc kubenswrapper[4735]: I0215 20:20:08.477926 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7vk2f" podStartSLOduration=3.641704992 podStartE2EDuration="1m3.47790683s" podCreationTimestamp="2026-02-15 20:19:05 +0000 UTC" firstStartedPulling="2026-02-15 20:19:08.115088637 +0000 UTC m=+155.981104260" lastFinishedPulling="2026-02-15 20:20:07.951290475 +0000 UTC m=+215.817306098" observedRunningTime="2026-02-15 20:20:08.469804222 +0000 UTC m=+216.335819855" watchObservedRunningTime="2026-02-15 20:20:08.47790683 +0000 UTC m=+216.343922453" Feb 15 20:20:08 crc kubenswrapper[4735]: I0215 20:20:08.526474 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4bxj7" podStartSLOduration=4.816916788 podStartE2EDuration="1m4.526455568s" podCreationTimestamp="2026-02-15 20:19:04 +0000 UTC" firstStartedPulling="2026-02-15 20:19:08.099895459 +0000 UTC m=+155.965911072" lastFinishedPulling="2026-02-15 20:20:07.809434229 +0000 UTC m=+215.675449852" observedRunningTime="2026-02-15 20:20:08.504224372 +0000 UTC m=+216.370239995" watchObservedRunningTime="2026-02-15 20:20:08.526455568 +0000 UTC m=+216.392471191" Feb 15 20:20:08 crc kubenswrapper[4735]: I0215 20:20:08.527969 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zwjpk" podStartSLOduration=3.930092552 podStartE2EDuration="1m2.52796194s" podCreationTimestamp="2026-02-15 20:19:06 +0000 UTC" firstStartedPulling="2026-02-15 20:19:09.199885653 +0000 UTC m=+157.065901276" lastFinishedPulling="2026-02-15 20:20:07.797755021 +0000 UTC m=+215.663770664" observedRunningTime="2026-02-15 20:20:08.522803865 +0000 UTC m=+216.388819498" watchObservedRunningTime="2026-02-15 20:20:08.52796194 +0000 UTC m=+216.393977563" Feb 15 20:20:08 crc kubenswrapper[4735]: I0215 20:20:08.892819 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f5c12d8-7152-4299-bcad-c59201575285" path="/var/lib/kubelet/pods/4f5c12d8-7152-4299-bcad-c59201575285/volumes" Feb 15 20:20:13 crc kubenswrapper[4735]: I0215 20:20:13.137203 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:20:13 crc kubenswrapper[4735]: I0215 20:20:13.137738 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:20:13 crc kubenswrapper[4735]: I0215 20:20:13.179018 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:20:13 crc kubenswrapper[4735]: I0215 20:20:13.197926 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lzl2l" podStartSLOduration=9.213470366 podStartE2EDuration="1m11.197908454s" podCreationTimestamp="2026-02-15 20:19:02 +0000 UTC" firstStartedPulling="2026-02-15 20:19:05.898689917 +0000 UTC m=+153.764705540" lastFinishedPulling="2026-02-15 20:20:07.883128005 +0000 UTC m=+215.749143628" observedRunningTime="2026-02-15 20:20:08.548163509 +0000 UTC m=+216.414179142" watchObservedRunningTime="2026-02-15 20:20:13.197908454 +0000 UTC m=+221.063924077" Feb 15 20:20:13 crc kubenswrapper[4735]: I0215 20:20:13.551334 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:20:15 crc kubenswrapper[4735]: I0215 20:20:15.338234 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:20:15 crc kubenswrapper[4735]: I0215 20:20:15.339461 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:20:15 crc kubenswrapper[4735]: I0215 20:20:15.403753 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:20:15 crc kubenswrapper[4735]: I0215 20:20:15.541528 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:20:16 crc kubenswrapper[4735]: I0215 20:20:16.313985 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:20:16 crc kubenswrapper[4735]: I0215 20:20:16.314277 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:20:16 crc kubenswrapper[4735]: I0215 20:20:16.349920 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:20:16 crc kubenswrapper[4735]: I0215 20:20:16.531400 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:20:16 crc kubenswrapper[4735]: I0215 20:20:16.658744 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:20:16 crc kubenswrapper[4735]: I0215 20:20:16.658785 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:20:16 crc kubenswrapper[4735]: I0215 20:20:16.703032 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:20:17 crc kubenswrapper[4735]: I0215 20:20:17.563421 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:20:18 crc kubenswrapper[4735]: I0215 20:20:18.585780 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bxj7"] Feb 15 20:20:18 crc kubenswrapper[4735]: I0215 20:20:18.586255 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4bxj7" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerName="registry-server" containerID="cri-o://4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77" gracePeriod=2 Feb 15 20:20:18 crc kubenswrapper[4735]: I0215 20:20:18.992761 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" podUID="9947c023-d2cc-4e2d-989c-fec1d09e4ca8" containerName="oauth-openshift" containerID="cri-o://56d37d29714a841c4bc005dbdd55124981688ee7a317302eb6bd82575d03c4a6" gracePeriod=15 Feb 15 20:20:18 crc kubenswrapper[4735]: I0215 20:20:18.997445 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.119418 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55xzs\" (UniqueName: \"kubernetes.io/projected/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-kube-api-access-55xzs\") pod \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.119554 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-utilities\") pod \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.119632 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-catalog-content\") pod \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\" (UID: \"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9\") " Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.123049 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-utilities" (OuterVolumeSpecName: "utilities") pod "853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" (UID: "853b0875-62d2-42fd-ad1a-c8fe3d16e8b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.137194 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-kube-api-access-55xzs" (OuterVolumeSpecName: "kube-api-access-55xzs") pod "853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" (UID: "853b0875-62d2-42fd-ad1a-c8fe3d16e8b9"). InnerVolumeSpecName "kube-api-access-55xzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.159161 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" (UID: "853b0875-62d2-42fd-ad1a-c8fe3d16e8b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.221363 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.221400 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.221929 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55xzs\" (UniqueName: \"kubernetes.io/projected/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9-kube-api-access-55xzs\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.509792 4735 generic.go:334] "Generic (PLEG): container finished" podID="9947c023-d2cc-4e2d-989c-fec1d09e4ca8" containerID="56d37d29714a841c4bc005dbdd55124981688ee7a317302eb6bd82575d03c4a6" exitCode=0 Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.509849 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" event={"ID":"9947c023-d2cc-4e2d-989c-fec1d09e4ca8","Type":"ContainerDied","Data":"56d37d29714a841c4bc005dbdd55124981688ee7a317302eb6bd82575d03c4a6"} Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.512402 4735 generic.go:334] "Generic (PLEG): container finished" podID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerID="4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77" exitCode=0 Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.512445 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bxj7" event={"ID":"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9","Type":"ContainerDied","Data":"4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77"} Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.512484 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bxj7" event={"ID":"853b0875-62d2-42fd-ad1a-c8fe3d16e8b9","Type":"ContainerDied","Data":"6338f5b6486ffef7db7f72ed1d5988f15f7252bc76fd8b8ccf2bac4a2210099f"} Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.512514 4735 scope.go:117] "RemoveContainer" containerID="4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.512863 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bxj7" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.534111 4735 scope.go:117] "RemoveContainer" containerID="68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.550129 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bxj7"] Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.553105 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bxj7"] Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.555845 4735 scope.go:117] "RemoveContainer" containerID="ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.569702 4735 scope.go:117] "RemoveContainer" containerID="4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77" Feb 15 20:20:19 crc kubenswrapper[4735]: E0215 20:20:19.570199 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77\": container with ID starting with 4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77 not found: ID does not exist" containerID="4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.570247 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77"} err="failed to get container status \"4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77\": rpc error: code = NotFound desc = could not find container \"4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77\": container with ID starting with 4fa8092917aa6775f89cca5b34fe8ab00f7445fe415e63b0adc68735bce31b77 not found: ID does not exist" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.570283 4735 scope.go:117] "RemoveContainer" containerID="68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300" Feb 15 20:20:19 crc kubenswrapper[4735]: E0215 20:20:19.570678 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300\": container with ID starting with 68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300 not found: ID does not exist" containerID="68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.570714 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300"} err="failed to get container status \"68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300\": rpc error: code = NotFound desc = could not find container \"68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300\": container with ID starting with 68332b6a873d23e5244046ddac1096bf09ae7edb5b9b3f7c193531136b988300 not found: ID does not exist" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.570744 4735 scope.go:117] "RemoveContainer" containerID="ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336" Feb 15 20:20:19 crc kubenswrapper[4735]: E0215 20:20:19.571013 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336\": container with ID starting with ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336 not found: ID does not exist" containerID="ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.571061 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336"} err="failed to get container status \"ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336\": rpc error: code = NotFound desc = could not find container \"ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336\": container with ID starting with ac57be157777eebe14de2f68558457485e8f182f365448ce1958eec6b3db5336 not found: ID does not exist" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.679841 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.679910 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.679979 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.680519 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.680575 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28" gracePeriod=600 Feb 15 20:20:19 crc kubenswrapper[4735]: I0215 20:20:19.905412 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.033253 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-idp-0-file-data\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.033304 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-cliconfig\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.033329 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-provider-selection\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.033357 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-serving-cert\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.033394 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-error\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.034225 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-ocp-branding-template\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.034258 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-session\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.034279 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-dir\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.034462 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-policies\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.034500 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-trusted-ca-bundle\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.034521 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-router-certs\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.034544 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-login\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.034579 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-service-ca\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.034599 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b885v\" (UniqueName: \"kubernetes.io/projected/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-kube-api-access-b885v\") pod \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\" (UID: \"9947c023-d2cc-4e2d-989c-fec1d09e4ca8\") " Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.036535 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.037028 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.037071 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.037549 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.041512 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.041886 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.042339 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.044561 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.044919 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.045926 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-kube-api-access-b885v" (OuterVolumeSpecName: "kube-api-access-b885v") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "kube-api-access-b885v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.054172 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.056035 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.056467 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.056696 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "9947c023-d2cc-4e2d-989c-fec1d09e4ca8" (UID: "9947c023-d2cc-4e2d-989c-fec1d09e4ca8"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136272 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136319 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136330 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136341 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136352 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136361 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136370 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136381 4735 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136391 4735 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136403 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136414 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136424 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136433 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.136442 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b885v\" (UniqueName: \"kubernetes.io/projected/9947c023-d2cc-4e2d-989c-fec1d09e4ca8-kube-api-access-b885v\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.518707 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28" exitCode=0 Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.518788 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28"} Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.520015 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" event={"ID":"9947c023-d2cc-4e2d-989c-fec1d09e4ca8","Type":"ContainerDied","Data":"192225447fdf282be618c96ec6af9a43a5e04e00941b397d9cbaef962138a803"} Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.520043 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6r89v" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.520083 4735 scope.go:117] "RemoveContainer" containerID="56d37d29714a841c4bc005dbdd55124981688ee7a317302eb6bd82575d03c4a6" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.542266 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r89v"] Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.545568 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r89v"] Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.785257 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zwjpk"] Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.785554 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zwjpk" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" containerName="registry-server" containerID="cri-o://21973c3c069f2bb78afe17f168d610d1c6eb8f69e3c64949a1366455012076b2" gracePeriod=2 Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.893681 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" path="/var/lib/kubelet/pods/853b0875-62d2-42fd-ad1a-c8fe3d16e8b9/volumes" Feb 15 20:20:20 crc kubenswrapper[4735]: I0215 20:20:20.894847 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9947c023-d2cc-4e2d-989c-fec1d09e4ca8" path="/var/lib/kubelet/pods/9947c023-d2cc-4e2d-989c-fec1d09e4ca8/volumes" Feb 15 20:20:21 crc kubenswrapper[4735]: I0215 20:20:21.530311 4735 generic.go:334] "Generic (PLEG): container finished" podID="aee39f47-6719-4de1-ab1c-5a8814345098" containerID="21973c3c069f2bb78afe17f168d610d1c6eb8f69e3c64949a1366455012076b2" exitCode=0 Feb 15 20:20:21 crc kubenswrapper[4735]: I0215 20:20:21.530427 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zwjpk" event={"ID":"aee39f47-6719-4de1-ab1c-5a8814345098","Type":"ContainerDied","Data":"21973c3c069f2bb78afe17f168d610d1c6eb8f69e3c64949a1366455012076b2"} Feb 15 20:20:21 crc kubenswrapper[4735]: I0215 20:20:21.535343 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"922ecbdcfc9b4bcb4d36975920d88c0963942c6e49ace39850b6e39e406cc403"} Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.026601 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.160463 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-utilities\") pod \"aee39f47-6719-4de1-ab1c-5a8814345098\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.160515 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-catalog-content\") pod \"aee39f47-6719-4de1-ab1c-5a8814345098\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.160607 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhbg8\" (UniqueName: \"kubernetes.io/projected/aee39f47-6719-4de1-ab1c-5a8814345098-kube-api-access-xhbg8\") pod \"aee39f47-6719-4de1-ab1c-5a8814345098\" (UID: \"aee39f47-6719-4de1-ab1c-5a8814345098\") " Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.162188 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-utilities" (OuterVolumeSpecName: "utilities") pod "aee39f47-6719-4de1-ab1c-5a8814345098" (UID: "aee39f47-6719-4de1-ab1c-5a8814345098"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.167101 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aee39f47-6719-4de1-ab1c-5a8814345098-kube-api-access-xhbg8" (OuterVolumeSpecName: "kube-api-access-xhbg8") pod "aee39f47-6719-4de1-ab1c-5a8814345098" (UID: "aee39f47-6719-4de1-ab1c-5a8814345098"). InnerVolumeSpecName "kube-api-access-xhbg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.262263 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhbg8\" (UniqueName: \"kubernetes.io/projected/aee39f47-6719-4de1-ab1c-5a8814345098-kube-api-access-xhbg8\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.262299 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.277253 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aee39f47-6719-4de1-ab1c-5a8814345098" (UID: "aee39f47-6719-4de1-ab1c-5a8814345098"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.363306 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee39f47-6719-4de1-ab1c-5a8814345098-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.544956 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zwjpk" event={"ID":"aee39f47-6719-4de1-ab1c-5a8814345098","Type":"ContainerDied","Data":"219a314d39da1d7be743247d50469ff73e5af97d6bae57a1b0e09cdabc0d3474"} Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.545020 4735 scope.go:117] "RemoveContainer" containerID="21973c3c069f2bb78afe17f168d610d1c6eb8f69e3c64949a1366455012076b2" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.545022 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zwjpk" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.567756 4735 scope.go:117] "RemoveContainer" containerID="5eef4918730b513010d0ee1f4e80e8f651f544920ad9294d46abc031ff0f3c1d" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.598462 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zwjpk"] Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.603011 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zwjpk"] Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.609886 4735 scope.go:117] "RemoveContainer" containerID="1340a1595134c3cdfb3920a15102720a3ab52af2f6d4f1e33152861e659d0dd1" Feb 15 20:20:22 crc kubenswrapper[4735]: I0215 20:20:22.896729 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" path="/var/lib/kubelet/pods/aee39f47-6719-4de1-ab1c-5a8814345098/volumes" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.754458 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7559487fb5-hnj49"] Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755229 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerName="extract-utilities" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755244 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerName="extract-utilities" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755256 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d90f00-7bbc-4358-9958-bf32d1bb967a" containerName="pruner" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755264 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d90f00-7bbc-4358-9958-bf32d1bb967a" containerName="pruner" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755278 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerName="extract-utilities" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755286 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerName="extract-utilities" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755297 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755304 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755313 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerName="extract-content" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755322 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerName="extract-content" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755335 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5c12d8-7152-4299-bcad-c59201575285" containerName="extract-utilities" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755342 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5c12d8-7152-4299-bcad-c59201575285" containerName="extract-utilities" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755357 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" containerName="extract-utilities" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755364 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" containerName="extract-utilities" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755373 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5c12d8-7152-4299-bcad-c59201575285" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755381 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5c12d8-7152-4299-bcad-c59201575285" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755393 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerName="extract-content" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755400 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerName="extract-content" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755412 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9947c023-d2cc-4e2d-989c-fec1d09e4ca8" containerName="oauth-openshift" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755419 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9947c023-d2cc-4e2d-989c-fec1d09e4ca8" containerName="oauth-openshift" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755431 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755440 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755451 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755458 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755469 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5c12d8-7152-4299-bcad-c59201575285" containerName="extract-content" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755476 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5c12d8-7152-4299-bcad-c59201575285" containerName="extract-content" Feb 15 20:20:27 crc kubenswrapper[4735]: E0215 20:20:27.755485 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" containerName="extract-content" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755492 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" containerName="extract-content" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755603 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f5c12d8-7152-4299-bcad-c59201575285" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755619 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9947c023-d2cc-4e2d-989c-fec1d09e4ca8" containerName="oauth-openshift" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755631 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="853b0875-62d2-42fd-ad1a-c8fe3d16e8b9" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755640 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf815092-d21d-4b8b-bbfa-fd11550ffb26" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755655 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee39f47-6719-4de1-ab1c-5a8814345098" containerName="registry-server" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.755664 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d90f00-7bbc-4358-9958-bf32d1bb967a" containerName="pruner" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.756136 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.760879 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.761378 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.765659 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.765726 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.765802 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.766130 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.766485 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.766580 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.766785 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.766904 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.767940 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.768118 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.796086 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.796545 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7559487fb5-hnj49"] Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.802707 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.813550 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932091 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-router-certs\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932144 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932174 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932196 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932213 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf4f9\" (UniqueName: \"kubernetes.io/projected/514823a9-b9bc-4544-aa3a-9b17933c9d63-kube-api-access-zf4f9\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932237 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-session\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932258 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-audit-policies\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932355 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932497 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-template-login\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932549 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932572 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-service-ca\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932609 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/514823a9-b9bc-4544-aa3a-9b17933c9d63-audit-dir\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932707 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-template-error\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:27 crc kubenswrapper[4735]: I0215 20:20:27.932827 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034280 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034350 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034374 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034460 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf4f9\" (UniqueName: \"kubernetes.io/projected/514823a9-b9bc-4544-aa3a-9b17933c9d63-kube-api-access-zf4f9\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034504 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-session\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034531 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-audit-policies\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034566 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034604 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-template-login\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034645 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034675 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-service-ca\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034710 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/514823a9-b9bc-4544-aa3a-9b17933c9d63-audit-dir\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034733 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-template-error\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034756 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.034798 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-router-certs\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.036366 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/514823a9-b9bc-4544-aa3a-9b17933c9d63-audit-dir\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.037518 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-audit-policies\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.038186 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.038381 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.038921 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-service-ca\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.042052 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.042285 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-template-error\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.042961 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.043263 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.044180 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-user-template-login\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.045700 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-router-certs\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.051363 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-session\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.055037 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/514823a9-b9bc-4544-aa3a-9b17933c9d63-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.064864 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf4f9\" (UniqueName: \"kubernetes.io/projected/514823a9-b9bc-4544-aa3a-9b17933c9d63-kube-api-access-zf4f9\") pod \"oauth-openshift-7559487fb5-hnj49\" (UID: \"514823a9-b9bc-4544-aa3a-9b17933c9d63\") " pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.079354 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.481067 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7559487fb5-hnj49"] Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.599801 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" event={"ID":"514823a9-b9bc-4544-aa3a-9b17933c9d63","Type":"ContainerStarted","Data":"4bb81124d973fc01a9a7393368ab02e83b7a680f57c445af634fb60dbc74e5d3"} Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.991489 4735 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.992618 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.992621 4735 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.993181 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de" gracePeriod=15 Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.993189 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e" gracePeriod=15 Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.993191 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3" gracePeriod=15 Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.993306 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad" gracePeriod=15 Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.993344 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392" gracePeriod=15 Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.993772 4735 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 15 20:20:28 crc kubenswrapper[4735]: E0215 20:20:28.993995 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994011 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 15 20:20:28 crc kubenswrapper[4735]: E0215 20:20:28.994024 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994039 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 15 20:20:28 crc kubenswrapper[4735]: E0215 20:20:28.994047 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994053 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 15 20:20:28 crc kubenswrapper[4735]: E0215 20:20:28.994064 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994069 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 15 20:20:28 crc kubenswrapper[4735]: E0215 20:20:28.994077 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994082 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 15 20:20:28 crc kubenswrapper[4735]: E0215 20:20:28.994091 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994096 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 15 20:20:28 crc kubenswrapper[4735]: E0215 20:20:28.994105 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994111 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994199 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994211 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994219 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994227 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994233 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 15 20:20:28 crc kubenswrapper[4735]: I0215 20:20:28.994241 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.035801 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.148137 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.148186 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.148219 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.148303 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.148329 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.148363 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.148422 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.148534 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250234 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250286 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250365 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250417 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250442 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250468 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250495 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250474 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250572 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250620 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250432 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250650 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250671 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250682 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.250700 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.335932 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:20:29 crc kubenswrapper[4735]: E0215 20:20:29.355659 4735 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.82:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189485098289c9fd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-15 20:20:29.354715645 +0000 UTC m=+237.220731278,LastTimestamp:2026-02-15 20:20:29.354715645 +0000 UTC m=+237.220731278,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.607716 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c322421a320f3fb8b21b06d40c6fd0f22cd5c49f521a21c401d154ffa9519b5d"} Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.608099 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a8f0774624512a5886d0c8691dd126b98512851413117ca6b2a9f6a8620c8e5a"} Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.608270 4735 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.608803 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.609907 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7559487fb5-hnj49_514823a9-b9bc-4544-aa3a-9b17933c9d63/oauth-openshift/0.log" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.609967 4735 generic.go:334] "Generic (PLEG): container finished" podID="514823a9-b9bc-4544-aa3a-9b17933c9d63" containerID="4571e18778b378f234c073c755fa5d0bbcdb08b03a9c04981009af3244301db5" exitCode=255 Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.610025 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" event={"ID":"514823a9-b9bc-4544-aa3a-9b17933c9d63","Type":"ContainerDied","Data":"4571e18778b378f234c073c755fa5d0bbcdb08b03a9c04981009af3244301db5"} Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.610856 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.611150 4735 scope.go:117] "RemoveContainer" containerID="4571e18778b378f234c073c755fa5d0bbcdb08b03a9c04981009af3244301db5" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.612683 4735 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.613032 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.615076 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.616559 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.617406 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e" exitCode=0 Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.617475 4735 scope.go:117] "RemoveContainer" containerID="45ba7080cf0df870e0792a5091f80b6651ecd9869d1b898ed434f0f1d55b227b" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.617585 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad" exitCode=0 Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.617597 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de" exitCode=0 Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.617603 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3" exitCode=2 Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.620005 4735 generic.go:334] "Generic (PLEG): container finished" podID="fa80450c-ee41-4f6f-813b-ba7556c751c0" containerID="4d0111dfd93f2c773dcd658b295fdde009d76f0cf2af8bbfc89cb9f45a5448ec" exitCode=0 Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.620030 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fa80450c-ee41-4f6f-813b-ba7556c751c0","Type":"ContainerDied","Data":"4d0111dfd93f2c773dcd658b295fdde009d76f0cf2af8bbfc89cb9f45a5448ec"} Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.620485 4735 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.620883 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.621276 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:29 crc kubenswrapper[4735]: I0215 20:20:29.621533 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.631735 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.638910 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7559487fb5-hnj49_514823a9-b9bc-4544-aa3a-9b17933c9d63/oauth-openshift/1.log" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.641211 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7559487fb5-hnj49_514823a9-b9bc-4544-aa3a-9b17933c9d63/oauth-openshift/0.log" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.641343 4735 generic.go:334] "Generic (PLEG): container finished" podID="514823a9-b9bc-4544-aa3a-9b17933c9d63" containerID="aa55fa5334d794e24396728251a8dfbd3e1bc536f4a531b6311ff224c322ae35" exitCode=255 Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.641506 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" event={"ID":"514823a9-b9bc-4544-aa3a-9b17933c9d63","Type":"ContainerDied","Data":"aa55fa5334d794e24396728251a8dfbd3e1bc536f4a531b6311ff224c322ae35"} Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.641672 4735 scope.go:117] "RemoveContainer" containerID="4571e18778b378f234c073c755fa5d0bbcdb08b03a9c04981009af3244301db5" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.643250 4735 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.643908 4735 scope.go:117] "RemoveContainer" containerID="aa55fa5334d794e24396728251a8dfbd3e1bc536f4a531b6311ff224c322ae35" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.644100 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:30 crc kubenswrapper[4735]: E0215 20:20:30.644627 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7559487fb5-hnj49_openshift-authentication(514823a9-b9bc-4544-aa3a-9b17933c9d63)\"" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.644642 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.647487 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.909128 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.910236 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.910726 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:30 crc kubenswrapper[4735]: I0215 20:20:30.911065 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.074904 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa80450c-ee41-4f6f-813b-ba7556c751c0-kube-api-access\") pod \"fa80450c-ee41-4f6f-813b-ba7556c751c0\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.075094 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-var-lock\") pod \"fa80450c-ee41-4f6f-813b-ba7556c751c0\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.075149 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-var-lock" (OuterVolumeSpecName: "var-lock") pod "fa80450c-ee41-4f6f-813b-ba7556c751c0" (UID: "fa80450c-ee41-4f6f-813b-ba7556c751c0"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.075216 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-kubelet-dir\") pod \"fa80450c-ee41-4f6f-813b-ba7556c751c0\" (UID: \"fa80450c-ee41-4f6f-813b-ba7556c751c0\") " Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.075246 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fa80450c-ee41-4f6f-813b-ba7556c751c0" (UID: "fa80450c-ee41-4f6f-813b-ba7556c751c0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.075683 4735 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-var-lock\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.075721 4735 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fa80450c-ee41-4f6f-813b-ba7556c751c0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.094332 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa80450c-ee41-4f6f-813b-ba7556c751c0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fa80450c-ee41-4f6f-813b-ba7556c751c0" (UID: "fa80450c-ee41-4f6f-813b-ba7556c751c0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.176722 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa80450c-ee41-4f6f-813b-ba7556c751c0-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.651214 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.652079 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392" exitCode=0 Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.654262 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fa80450c-ee41-4f6f-813b-ba7556c751c0","Type":"ContainerDied","Data":"a21272bfc8258468f2df11a28a6c17806b67fa0cf574d2c2782084eb6d6071f0"} Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.654302 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a21272bfc8258468f2df11a28a6c17806b67fa0cf574d2c2782084eb6d6071f0" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.654323 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.655681 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7559487fb5-hnj49_514823a9-b9bc-4544-aa3a-9b17933c9d63/oauth-openshift/1.log" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.656154 4735 scope.go:117] "RemoveContainer" containerID="aa55fa5334d794e24396728251a8dfbd3e1bc536f4a531b6311ff224c322ae35" Feb 15 20:20:31 crc kubenswrapper[4735]: E0215 20:20:31.656416 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7559487fb5-hnj49_openshift-authentication(514823a9-b9bc-4544-aa3a-9b17933c9d63)\"" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.656493 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.656877 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.657127 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.674640 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.675215 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.675574 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.882537 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.883533 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.884341 4735 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.884665 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.885006 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.885489 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.988412 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.988542 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.988572 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.988533 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.988566 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.988709 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.988871 4735 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.988892 4735 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:31 crc kubenswrapper[4735]: I0215 20:20:31.988903 4735 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.663311 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.664019 4735 scope.go:117] "RemoveContainer" containerID="e87285f17c592173376d4ea72b9894094c78703650918fc0a7b68cfecce4b72e" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.664094 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.681703 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.683069 4735 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.683970 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.684359 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.685784 4735 scope.go:117] "RemoveContainer" containerID="9f760d06cc7a882309ebdc2142b7e7cfe6023a2926973210e598d39b10f2dfad" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.697641 4735 scope.go:117] "RemoveContainer" containerID="0674afc3a60c6d6069e2ec2015d7c7e96eed18aa424df4e3a403b2ec9a9533de" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.714286 4735 scope.go:117] "RemoveContainer" containerID="41aac7dbde1ca895bdb834a78e22702602a038d805ba8a7882225e935d73c1f3" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.733667 4735 scope.go:117] "RemoveContainer" containerID="7fbc893ea396a9bfbd401c4eaf53f5dad8bd1cd9fdc772c2318764523b5bc392" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.756881 4735 scope.go:117] "RemoveContainer" containerID="fc045441d319ed45248fc79c132f3a98659654ed9cbf072c4b88ea2ace3e0934" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.896409 4735 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.896866 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.897655 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.898179 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:32 crc kubenswrapper[4735]: I0215 20:20:32.901406 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 15 20:20:33 crc kubenswrapper[4735]: E0215 20:20:33.167819 4735 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:33 crc kubenswrapper[4735]: E0215 20:20:33.168111 4735 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:33 crc kubenswrapper[4735]: E0215 20:20:33.168368 4735 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:33 crc kubenswrapper[4735]: E0215 20:20:33.168616 4735 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:33 crc kubenswrapper[4735]: E0215 20:20:33.168861 4735 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:33 crc kubenswrapper[4735]: I0215 20:20:33.168893 4735 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 15 20:20:33 crc kubenswrapper[4735]: E0215 20:20:33.169179 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="200ms" Feb 15 20:20:33 crc kubenswrapper[4735]: E0215 20:20:33.369855 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="400ms" Feb 15 20:20:33 crc kubenswrapper[4735]: E0215 20:20:33.770853 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="800ms" Feb 15 20:20:34 crc kubenswrapper[4735]: E0215 20:20:34.573566 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="1.6s" Feb 15 20:20:34 crc kubenswrapper[4735]: E0215 20:20:34.591874 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:20:34Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:20:34Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:20:34Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-15T20:20:34Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:34 crc kubenswrapper[4735]: E0215 20:20:34.592243 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:34 crc kubenswrapper[4735]: E0215 20:20:34.592662 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:34 crc kubenswrapper[4735]: E0215 20:20:34.593162 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:34 crc kubenswrapper[4735]: E0215 20:20:34.593421 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:34 crc kubenswrapper[4735]: E0215 20:20:34.593443 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 15 20:20:36 crc kubenswrapper[4735]: E0215 20:20:36.175710 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="3.2s" Feb 15 20:20:37 crc kubenswrapper[4735]: E0215 20:20:37.165313 4735 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.82:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189485098289c9fd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-15 20:20:29.354715645 +0000 UTC m=+237.220731278,LastTimestamp:2026-02-15 20:20:29.354715645 +0000 UTC m=+237.220731278,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 15 20:20:38 crc kubenswrapper[4735]: I0215 20:20:38.080073 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:38 crc kubenswrapper[4735]: I0215 20:20:38.080126 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:38 crc kubenswrapper[4735]: I0215 20:20:38.080802 4735 scope.go:117] "RemoveContainer" containerID="aa55fa5334d794e24396728251a8dfbd3e1bc536f4a531b6311ff224c322ae35" Feb 15 20:20:38 crc kubenswrapper[4735]: E0215 20:20:38.081047 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7559487fb5-hnj49_openshift-authentication(514823a9-b9bc-4544-aa3a-9b17933c9d63)\"" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" Feb 15 20:20:39 crc kubenswrapper[4735]: E0215 20:20:39.376822 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.82:6443: connect: connection refused" interval="6.4s" Feb 15 20:20:40 crc kubenswrapper[4735]: I0215 20:20:40.886392 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:40 crc kubenswrapper[4735]: I0215 20:20:40.888529 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:40 crc kubenswrapper[4735]: I0215 20:20:40.889062 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:40 crc kubenswrapper[4735]: I0215 20:20:40.890096 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:40 crc kubenswrapper[4735]: I0215 20:20:40.911523 4735 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:40 crc kubenswrapper[4735]: I0215 20:20:40.911792 4735 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:40 crc kubenswrapper[4735]: E0215 20:20:40.912562 4735 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:40 crc kubenswrapper[4735]: I0215 20:20:40.913344 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:40 crc kubenswrapper[4735]: W0215 20:20:40.938361 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-b223057785eb6d88e79264196233570c4709aa6fab5c59615c9c062c245176ca WatchSource:0}: Error finding container b223057785eb6d88e79264196233570c4709aa6fab5c59615c9c062c245176ca: Status 404 returned error can't find the container with id b223057785eb6d88e79264196233570c4709aa6fab5c59615c9c062c245176ca Feb 15 20:20:41 crc kubenswrapper[4735]: I0215 20:20:41.718120 4735 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="565a0defc6cce8f8f02da40be9263b72e3edf309a6dd60c0888c177700f1297f" exitCode=0 Feb 15 20:20:41 crc kubenswrapper[4735]: I0215 20:20:41.718244 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"565a0defc6cce8f8f02da40be9263b72e3edf309a6dd60c0888c177700f1297f"} Feb 15 20:20:41 crc kubenswrapper[4735]: I0215 20:20:41.718443 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b223057785eb6d88e79264196233570c4709aa6fab5c59615c9c062c245176ca"} Feb 15 20:20:41 crc kubenswrapper[4735]: I0215 20:20:41.718727 4735 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:41 crc kubenswrapper[4735]: I0215 20:20:41.718743 4735 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:41 crc kubenswrapper[4735]: E0215 20:20:41.719293 4735 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:41 crc kubenswrapper[4735]: I0215 20:20:41.719370 4735 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:41 crc kubenswrapper[4735]: I0215 20:20:41.720013 4735 status_manager.go:851] "Failed to get status for pod" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:41 crc kubenswrapper[4735]: I0215 20:20:41.720441 4735 status_manager.go:851] "Failed to get status for pod" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7559487fb5-hnj49\": dial tcp 38.102.83.82:6443: connect: connection refused" Feb 15 20:20:42 crc kubenswrapper[4735]: I0215 20:20:42.741443 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 15 20:20:42 crc kubenswrapper[4735]: I0215 20:20:42.741488 4735 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063" exitCode=1 Feb 15 20:20:42 crc kubenswrapper[4735]: I0215 20:20:42.741544 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063"} Feb 15 20:20:42 crc kubenswrapper[4735]: I0215 20:20:42.741905 4735 scope.go:117] "RemoveContainer" containerID="e579ea97843e259460cd9d6ea61e66cb7d2d4c7b5e59df55a3e68f7d6fe92063" Feb 15 20:20:42 crc kubenswrapper[4735]: I0215 20:20:42.746107 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fdc8f8710260b7abcbe24ed2151bc91f07e124b0c2fb66740a173eaaa0aa8487"} Feb 15 20:20:42 crc kubenswrapper[4735]: I0215 20:20:42.746144 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"64031ea2b6ce174a4fde4e3a4ab2d51fe6287f1c08389caee099da4f3317c118"} Feb 15 20:20:42 crc kubenswrapper[4735]: I0215 20:20:42.746154 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"108e0a97402f66c44cc5e67f245d8f662b50cc781cc3c4c74cdf21858c7e9524"} Feb 15 20:20:42 crc kubenswrapper[4735]: I0215 20:20:42.855418 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:20:43 crc kubenswrapper[4735]: I0215 20:20:43.752516 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"89d9171825a88b1ba380c0b78522139b272648048569555673e70084bf701af4"} Feb 15 20:20:43 crc kubenswrapper[4735]: I0215 20:20:43.752755 4735 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:43 crc kubenswrapper[4735]: I0215 20:20:43.752783 4735 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:43 crc kubenswrapper[4735]: I0215 20:20:43.752783 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cf16c61ae76de243db78cc2ade5e9420ab484df7b71542a4033eaf00667e6349"} Feb 15 20:20:43 crc kubenswrapper[4735]: I0215 20:20:43.752852 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:43 crc kubenswrapper[4735]: I0215 20:20:43.755359 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 15 20:20:43 crc kubenswrapper[4735]: I0215 20:20:43.755468 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4238ab5d16945a52c301a4c1353967cf6a1db8e5fe74cb9ee9851d7dd3da5b1c"} Feb 15 20:20:44 crc kubenswrapper[4735]: I0215 20:20:44.344876 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:20:45 crc kubenswrapper[4735]: I0215 20:20:45.914388 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:45 crc kubenswrapper[4735]: I0215 20:20:45.914671 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:45 crc kubenswrapper[4735]: I0215 20:20:45.919348 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:48 crc kubenswrapper[4735]: I0215 20:20:48.760736 4735 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:49 crc kubenswrapper[4735]: I0215 20:20:49.786492 4735 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:49 crc kubenswrapper[4735]: I0215 20:20:49.787113 4735 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:49 crc kubenswrapper[4735]: I0215 20:20:49.790481 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:20:49 crc kubenswrapper[4735]: I0215 20:20:49.792964 4735 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="11b837dc-01e8-4155-9f24-7ccbd227de1c" Feb 15 20:20:50 crc kubenswrapper[4735]: I0215 20:20:50.792241 4735 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:50 crc kubenswrapper[4735]: I0215 20:20:50.792289 4735 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e0a1cc0b-b9b2-436b-8d7b-70b9018d82b7" Feb 15 20:20:51 crc kubenswrapper[4735]: I0215 20:20:51.887099 4735 scope.go:117] "RemoveContainer" containerID="aa55fa5334d794e24396728251a8dfbd3e1bc536f4a531b6311ff224c322ae35" Feb 15 20:20:52 crc kubenswrapper[4735]: I0215 20:20:52.803092 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7559487fb5-hnj49_514823a9-b9bc-4544-aa3a-9b17933c9d63/oauth-openshift/2.log" Feb 15 20:20:52 crc kubenswrapper[4735]: I0215 20:20:52.803856 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7559487fb5-hnj49_514823a9-b9bc-4544-aa3a-9b17933c9d63/oauth-openshift/1.log" Feb 15 20:20:52 crc kubenswrapper[4735]: I0215 20:20:52.803897 4735 generic.go:334] "Generic (PLEG): container finished" podID="514823a9-b9bc-4544-aa3a-9b17933c9d63" containerID="f5714082bfc386cad6267f4b9def041bb236f37c9ffe0852f35d395eb811a75c" exitCode=255 Feb 15 20:20:52 crc kubenswrapper[4735]: I0215 20:20:52.803930 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" event={"ID":"514823a9-b9bc-4544-aa3a-9b17933c9d63","Type":"ContainerDied","Data":"f5714082bfc386cad6267f4b9def041bb236f37c9ffe0852f35d395eb811a75c"} Feb 15 20:20:52 crc kubenswrapper[4735]: I0215 20:20:52.804014 4735 scope.go:117] "RemoveContainer" containerID="aa55fa5334d794e24396728251a8dfbd3e1bc536f4a531b6311ff224c322ae35" Feb 15 20:20:52 crc kubenswrapper[4735]: I0215 20:20:52.804556 4735 scope.go:117] "RemoveContainer" containerID="f5714082bfc386cad6267f4b9def041bb236f37c9ffe0852f35d395eb811a75c" Feb 15 20:20:52 crc kubenswrapper[4735]: E0215 20:20:52.804741 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-7559487fb5-hnj49_openshift-authentication(514823a9-b9bc-4544-aa3a-9b17933c9d63)\"" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" Feb 15 20:20:52 crc kubenswrapper[4735]: I0215 20:20:52.855391 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:20:52 crc kubenswrapper[4735]: I0215 20:20:52.859689 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:20:52 crc kubenswrapper[4735]: I0215 20:20:52.900378 4735 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="11b837dc-01e8-4155-9f24-7ccbd227de1c" Feb 15 20:20:53 crc kubenswrapper[4735]: I0215 20:20:53.811727 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7559487fb5-hnj49_514823a9-b9bc-4544-aa3a-9b17933c9d63/oauth-openshift/2.log" Feb 15 20:20:53 crc kubenswrapper[4735]: I0215 20:20:53.819381 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 15 20:20:58 crc kubenswrapper[4735]: I0215 20:20:58.079545 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:58 crc kubenswrapper[4735]: I0215 20:20:58.080253 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:20:58 crc kubenswrapper[4735]: I0215 20:20:58.081095 4735 scope.go:117] "RemoveContainer" containerID="f5714082bfc386cad6267f4b9def041bb236f37c9ffe0852f35d395eb811a75c" Feb 15 20:20:58 crc kubenswrapper[4735]: E0215 20:20:58.081399 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-7559487fb5-hnj49_openshift-authentication(514823a9-b9bc-4544-aa3a-9b17933c9d63)\"" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" Feb 15 20:20:59 crc kubenswrapper[4735]: I0215 20:20:59.599829 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 15 20:20:59 crc kubenswrapper[4735]: I0215 20:20:59.715315 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 15 20:21:00 crc kubenswrapper[4735]: I0215 20:21:00.410993 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 15 20:21:00 crc kubenswrapper[4735]: I0215 20:21:00.612694 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 15 20:21:00 crc kubenswrapper[4735]: I0215 20:21:00.784152 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 15 20:21:00 crc kubenswrapper[4735]: I0215 20:21:00.829653 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 15 20:21:00 crc kubenswrapper[4735]: I0215 20:21:00.829653 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.158851 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.239572 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.322877 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.595706 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.657438 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.672014 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.793919 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.794151 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.871912 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 15 20:21:01 crc kubenswrapper[4735]: I0215 20:21:01.880100 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.193626 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.227759 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.278619 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.323047 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.371768 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.377766 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.439556 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.606525 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.630401 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.756846 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.766730 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.813398 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.813996 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.836691 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.847910 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.931890 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 15 20:21:02 crc kubenswrapper[4735]: I0215 20:21:02.953508 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.071604 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.153812 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.170624 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.273276 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.344837 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.379455 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.394660 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.439705 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.451399 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.456336 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.490244 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.510542 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.561525 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.594721 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.602883 4735 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.618260 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.647270 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.698060 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.791672 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.794247 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.794858 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.885066 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.886845 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.952717 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 15 20:21:03 crc kubenswrapper[4735]: I0215 20:21:03.988890 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.076476 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.120656 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.168727 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.336010 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.423200 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.531505 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.640801 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.757585 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.780069 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.808293 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.822544 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.881910 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.887936 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.890690 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 15 20:21:04 crc kubenswrapper[4735]: I0215 20:21:04.982129 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.080871 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.113563 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.113829 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.202909 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.208682 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.240139 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.261278 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.298788 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.315637 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.325055 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.333468 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.343712 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.392690 4735 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.410556 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.410584 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.524384 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.601397 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.660702 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.698881 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.816821 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.823703 4735 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.846298 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.883222 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.915845 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.945722 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 15 20:21:05 crc kubenswrapper[4735]: I0215 20:21:05.962162 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.104669 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.127834 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.139516 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.143369 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.187305 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.354077 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.490287 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.539482 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.543247 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.668246 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 15 20:21:06 crc kubenswrapper[4735]: I0215 20:21:06.902590 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.012850 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.109010 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.161314 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.234984 4735 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.381718 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.543510 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.601669 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.740590 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.820887 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.840124 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.881212 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.896297 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.936786 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.938862 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.971315 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 15 20:21:07 crc kubenswrapper[4735]: I0215 20:21:07.985094 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.092727 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.104643 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.111653 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.206832 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.220903 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.229349 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.229992 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.347836 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.483268 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.515781 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.609440 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.656318 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.791430 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.836771 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 15 20:21:08 crc kubenswrapper[4735]: I0215 20:21:08.872683 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.093760 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.106487 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.118676 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.134339 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.143293 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.172572 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.188140 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.260203 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.262214 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.269067 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.296925 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.298494 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.326112 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.377458 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.423444 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.462619 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.514798 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.569823 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.614088 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.637224 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.734531 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.756654 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.762441 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.826600 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.857230 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.912869 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.969251 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 15 20:21:09 crc kubenswrapper[4735]: I0215 20:21:09.996594 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.117640 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.426502 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.540527 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.588268 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.626983 4735 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.635685 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.636275 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.684030 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.741629 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.781258 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.786962 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.877657 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.889591 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.938155 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 15 20:21:10 crc kubenswrapper[4735]: I0215 20:21:10.944547 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.237684 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.271683 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.295862 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.451060 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.517771 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.527337 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.548826 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.608981 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.628890 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.666097 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.694834 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.830563 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.834976 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.858070 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.886495 4735 scope.go:117] "RemoveContainer" containerID="f5714082bfc386cad6267f4b9def041bb236f37c9ffe0852f35d395eb811a75c" Feb 15 20:21:11 crc kubenswrapper[4735]: E0215 20:21:11.886859 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-7559487fb5-hnj49_openshift-authentication(514823a9-b9bc-4544-aa3a-9b17933c9d63)\"" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" podUID="514823a9-b9bc-4544-aa3a-9b17933c9d63" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.895670 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 15 20:21:11 crc kubenswrapper[4735]: I0215 20:21:11.999443 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.024643 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.093826 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.094497 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.228292 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.238432 4735 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.239610 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=43.239583495 podStartE2EDuration="43.239583495s" podCreationTimestamp="2026-02-15 20:20:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:20:48.602938365 +0000 UTC m=+256.468954008" watchObservedRunningTime="2026-02-15 20:21:12.239583495 +0000 UTC m=+280.105599158" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.245589 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.245642 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.251368 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.264928 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=24.264903787 podStartE2EDuration="24.264903787s" podCreationTimestamp="2026-02-15 20:20:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:21:12.261876062 +0000 UTC m=+280.127891685" watchObservedRunningTime="2026-02-15 20:21:12.264903787 +0000 UTC m=+280.130919430" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.391058 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.391396 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.402180 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.616165 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.646378 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.751558 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.848634 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.848725 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 15 20:21:12 crc kubenswrapper[4735]: I0215 20:21:12.963831 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.056554 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.075650 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.083724 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.092236 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.093346 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.179487 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.228564 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.349589 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.366478 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.367814 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.660550 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 15 20:21:13 crc kubenswrapper[4735]: I0215 20:21:13.862918 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.022982 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.052454 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.073123 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.165333 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.190538 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.304286 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.399828 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.733525 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.811033 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.859725 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 15 20:21:14 crc kubenswrapper[4735]: I0215 20:21:14.983271 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 15 20:21:15 crc kubenswrapper[4735]: I0215 20:21:15.360755 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 15 20:21:15 crc kubenswrapper[4735]: I0215 20:21:15.514832 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 15 20:21:15 crc kubenswrapper[4735]: I0215 20:21:15.561881 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 15 20:21:16 crc kubenswrapper[4735]: I0215 20:21:16.036834 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 15 20:21:16 crc kubenswrapper[4735]: I0215 20:21:16.317462 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 15 20:21:16 crc kubenswrapper[4735]: I0215 20:21:16.758578 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 15 20:21:16 crc kubenswrapper[4735]: I0215 20:21:16.936233 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 15 20:21:17 crc kubenswrapper[4735]: I0215 20:21:17.210273 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 15 20:21:17 crc kubenswrapper[4735]: I0215 20:21:17.273272 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 15 20:21:17 crc kubenswrapper[4735]: I0215 20:21:17.508258 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 15 20:21:17 crc kubenswrapper[4735]: I0215 20:21:17.564184 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 15 20:21:17 crc kubenswrapper[4735]: I0215 20:21:17.627932 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 15 20:21:17 crc kubenswrapper[4735]: I0215 20:21:17.898250 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 15 20:21:18 crc kubenswrapper[4735]: I0215 20:21:18.505611 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 15 20:21:22 crc kubenswrapper[4735]: I0215 20:21:22.626071 4735 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 15 20:21:22 crc kubenswrapper[4735]: I0215 20:21:22.626561 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://c322421a320f3fb8b21b06d40c6fd0f22cd5c49f521a21c401d154ffa9519b5d" gracePeriod=5 Feb 15 20:21:24 crc kubenswrapper[4735]: I0215 20:21:24.887329 4735 scope.go:117] "RemoveContainer" containerID="f5714082bfc386cad6267f4b9def041bb236f37c9ffe0852f35d395eb811a75c" Feb 15 20:21:25 crc kubenswrapper[4735]: I0215 20:21:25.986800 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7559487fb5-hnj49_514823a9-b9bc-4544-aa3a-9b17933c9d63/oauth-openshift/2.log" Feb 15 20:21:25 crc kubenswrapper[4735]: I0215 20:21:25.987157 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" event={"ID":"514823a9-b9bc-4544-aa3a-9b17933c9d63","Type":"ContainerStarted","Data":"94c57426b953553463d81f0d1736808e55c517bdbb4ef95370b2840ce4dd9755"} Feb 15 20:21:25 crc kubenswrapper[4735]: I0215 20:21:25.987614 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:21:25 crc kubenswrapper[4735]: I0215 20:21:25.995937 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" Feb 15 20:21:26 crc kubenswrapper[4735]: I0215 20:21:26.046646 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7559487fb5-hnj49" podStartSLOduration=93.046626988 podStartE2EDuration="1m33.046626988s" podCreationTimestamp="2026-02-15 20:19:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:21:26.023758455 +0000 UTC m=+293.889774078" watchObservedRunningTime="2026-02-15 20:21:26.046626988 +0000 UTC m=+293.912642621" Feb 15 20:21:27 crc kubenswrapper[4735]: I0215 20:21:27.997747 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 15 20:21:27 crc kubenswrapper[4735]: I0215 20:21:27.998032 4735 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="c322421a320f3fb8b21b06d40c6fd0f22cd5c49f521a21c401d154ffa9519b5d" exitCode=137 Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.213477 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.213608 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316296 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316376 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316406 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316429 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316489 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316508 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316531 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316560 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316638 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316934 4735 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316977 4735 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.316991 4735 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.317002 4735 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.325162 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.417733 4735 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.892274 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.892715 4735 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.904984 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.905015 4735 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b891358c-0260-444e-8dc9-34efcf3b4604" Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.906110 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 15 20:21:28 crc kubenswrapper[4735]: I0215 20:21:28.906131 4735 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b891358c-0260-444e-8dc9-34efcf3b4604" Feb 15 20:21:29 crc kubenswrapper[4735]: I0215 20:21:29.008764 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 15 20:21:29 crc kubenswrapper[4735]: I0215 20:21:29.008834 4735 scope.go:117] "RemoveContainer" containerID="c322421a320f3fb8b21b06d40c6fd0f22cd5c49f521a21c401d154ffa9519b5d" Feb 15 20:21:29 crc kubenswrapper[4735]: I0215 20:21:29.009040 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 15 20:21:32 crc kubenswrapper[4735]: I0215 20:21:32.648540 4735 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 15 20:21:35 crc kubenswrapper[4735]: I0215 20:21:35.043103 4735 generic.go:334] "Generic (PLEG): container finished" podID="1626b474-7e91-4cde-bab7-a4db753478dd" containerID="be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835" exitCode=0 Feb 15 20:21:35 crc kubenswrapper[4735]: I0215 20:21:35.043209 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" event={"ID":"1626b474-7e91-4cde-bab7-a4db753478dd","Type":"ContainerDied","Data":"be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835"} Feb 15 20:21:35 crc kubenswrapper[4735]: I0215 20:21:35.044195 4735 scope.go:117] "RemoveContainer" containerID="be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835" Feb 15 20:21:35 crc kubenswrapper[4735]: I0215 20:21:35.571197 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:21:35 crc kubenswrapper[4735]: I0215 20:21:35.571248 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:21:36 crc kubenswrapper[4735]: I0215 20:21:36.052433 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" event={"ID":"1626b474-7e91-4cde-bab7-a4db753478dd","Type":"ContainerStarted","Data":"85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261"} Feb 15 20:21:37 crc kubenswrapper[4735]: I0215 20:21:37.056336 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:21:37 crc kubenswrapper[4735]: I0215 20:21:37.058667 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.279955 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jfpmq"] Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.280606 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" podUID="36865289-7421-41dd-8434-fb2b31a3b974" containerName="controller-manager" containerID="cri-o://1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428" gracePeriod=30 Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.379409 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh"] Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.379789 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" podUID="3fb3271e-4d33-4ef0-b269-ca69f698fb66" containerName="route-controller-manager" containerID="cri-o://2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8" gracePeriod=30 Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.613432 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.691628 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-config\") pod \"36865289-7421-41dd-8434-fb2b31a3b974\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.691707 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-proxy-ca-bundles\") pod \"36865289-7421-41dd-8434-fb2b31a3b974\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.691751 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36865289-7421-41dd-8434-fb2b31a3b974-serving-cert\") pod \"36865289-7421-41dd-8434-fb2b31a3b974\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.691875 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm6c6\" (UniqueName: \"kubernetes.io/projected/36865289-7421-41dd-8434-fb2b31a3b974-kube-api-access-wm6c6\") pod \"36865289-7421-41dd-8434-fb2b31a3b974\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.691911 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-client-ca\") pod \"36865289-7421-41dd-8434-fb2b31a3b974\" (UID: \"36865289-7421-41dd-8434-fb2b31a3b974\") " Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.692929 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-client-ca" (OuterVolumeSpecName: "client-ca") pod "36865289-7421-41dd-8434-fb2b31a3b974" (UID: "36865289-7421-41dd-8434-fb2b31a3b974"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.693613 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-config" (OuterVolumeSpecName: "config") pod "36865289-7421-41dd-8434-fb2b31a3b974" (UID: "36865289-7421-41dd-8434-fb2b31a3b974"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.694701 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "36865289-7421-41dd-8434-fb2b31a3b974" (UID: "36865289-7421-41dd-8434-fb2b31a3b974"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.715234 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36865289-7421-41dd-8434-fb2b31a3b974-kube-api-access-wm6c6" (OuterVolumeSpecName: "kube-api-access-wm6c6") pod "36865289-7421-41dd-8434-fb2b31a3b974" (UID: "36865289-7421-41dd-8434-fb2b31a3b974"). InnerVolumeSpecName "kube-api-access-wm6c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.716366 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36865289-7421-41dd-8434-fb2b31a3b974-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "36865289-7421-41dd-8434-fb2b31a3b974" (UID: "36865289-7421-41dd-8434-fb2b31a3b974"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.761361 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.793029 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.793066 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.793080 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36865289-7421-41dd-8434-fb2b31a3b974-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.793093 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm6c6\" (UniqueName: \"kubernetes.io/projected/36865289-7421-41dd-8434-fb2b31a3b974-kube-api-access-wm6c6\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.793103 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36865289-7421-41dd-8434-fb2b31a3b974-client-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.894351 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fb3271e-4d33-4ef0-b269-ca69f698fb66-serving-cert\") pod \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.894410 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-client-ca\") pod \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.894449 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpkt9\" (UniqueName: \"kubernetes.io/projected/3fb3271e-4d33-4ef0-b269-ca69f698fb66-kube-api-access-fpkt9\") pod \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.894523 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-config\") pod \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\" (UID: \"3fb3271e-4d33-4ef0-b269-ca69f698fb66\") " Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.895107 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-client-ca" (OuterVolumeSpecName: "client-ca") pod "3fb3271e-4d33-4ef0-b269-ca69f698fb66" (UID: "3fb3271e-4d33-4ef0-b269-ca69f698fb66"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.895165 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-config" (OuterVolumeSpecName: "config") pod "3fb3271e-4d33-4ef0-b269-ca69f698fb66" (UID: "3fb3271e-4d33-4ef0-b269-ca69f698fb66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.897835 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fb3271e-4d33-4ef0-b269-ca69f698fb66-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3fb3271e-4d33-4ef0-b269-ca69f698fb66" (UID: "3fb3271e-4d33-4ef0-b269-ca69f698fb66"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.898001 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb3271e-4d33-4ef0-b269-ca69f698fb66-kube-api-access-fpkt9" (OuterVolumeSpecName: "kube-api-access-fpkt9") pod "3fb3271e-4d33-4ef0-b269-ca69f698fb66" (UID: "3fb3271e-4d33-4ef0-b269-ca69f698fb66"). InnerVolumeSpecName "kube-api-access-fpkt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.995439 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.995471 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fb3271e-4d33-4ef0-b269-ca69f698fb66-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.995480 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fb3271e-4d33-4ef0-b269-ca69f698fb66-client-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:21 crc kubenswrapper[4735]: I0215 20:22:21.995488 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpkt9\" (UniqueName: \"kubernetes.io/projected/3fb3271e-4d33-4ef0-b269-ca69f698fb66-kube-api-access-fpkt9\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.308347 4735 generic.go:334] "Generic (PLEG): container finished" podID="36865289-7421-41dd-8434-fb2b31a3b974" containerID="1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428" exitCode=0 Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.308394 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.308433 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" event={"ID":"36865289-7421-41dd-8434-fb2b31a3b974","Type":"ContainerDied","Data":"1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428"} Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.309154 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jfpmq" event={"ID":"36865289-7421-41dd-8434-fb2b31a3b974","Type":"ContainerDied","Data":"84e08b76e7809727f99d9c58c0aad2562c1dc8394a82860ff101b489d6328037"} Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.309199 4735 scope.go:117] "RemoveContainer" containerID="1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.313705 4735 generic.go:334] "Generic (PLEG): container finished" podID="3fb3271e-4d33-4ef0-b269-ca69f698fb66" containerID="2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8" exitCode=0 Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.313752 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" event={"ID":"3fb3271e-4d33-4ef0-b269-ca69f698fb66","Type":"ContainerDied","Data":"2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8"} Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.313783 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" event={"ID":"3fb3271e-4d33-4ef0-b269-ca69f698fb66","Type":"ContainerDied","Data":"07f83cd951594661fd420167099f0926e8c47b3017b98a3b3498a4bff3398538"} Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.313855 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.331748 4735 scope.go:117] "RemoveContainer" containerID="1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428" Feb 15 20:22:22 crc kubenswrapper[4735]: E0215 20:22:22.332918 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428\": container with ID starting with 1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428 not found: ID does not exist" containerID="1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.333064 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428"} err="failed to get container status \"1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428\": rpc error: code = NotFound desc = could not find container \"1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428\": container with ID starting with 1af7ca3605a9d641dee886f71ef086c431d1b40ec8b840af03d8fdee54567428 not found: ID does not exist" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.333175 4735 scope.go:117] "RemoveContainer" containerID="2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.347653 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh"] Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.352191 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tgtkh"] Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.364937 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jfpmq"] Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.370188 4735 scope.go:117] "RemoveContainer" containerID="2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8" Feb 15 20:22:22 crc kubenswrapper[4735]: E0215 20:22:22.370729 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8\": container with ID starting with 2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8 not found: ID does not exist" containerID="2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.370769 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8"} err="failed to get container status \"2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8\": rpc error: code = NotFound desc = could not find container \"2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8\": container with ID starting with 2f854a73e192f80155f1a3ee1de384141357d3f7f42b9fee373d35b8401329b8 not found: ID does not exist" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.372259 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jfpmq"] Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825264 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl"] Feb 15 20:22:22 crc kubenswrapper[4735]: E0215 20:22:22.825461 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" containerName="installer" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825473 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" containerName="installer" Feb 15 20:22:22 crc kubenswrapper[4735]: E0215 20:22:22.825482 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36865289-7421-41dd-8434-fb2b31a3b974" containerName="controller-manager" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825488 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="36865289-7421-41dd-8434-fb2b31a3b974" containerName="controller-manager" Feb 15 20:22:22 crc kubenswrapper[4735]: E0215 20:22:22.825497 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825503 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 15 20:22:22 crc kubenswrapper[4735]: E0215 20:22:22.825512 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb3271e-4d33-4ef0-b269-ca69f698fb66" containerName="route-controller-manager" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825518 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb3271e-4d33-4ef0-b269-ca69f698fb66" containerName="route-controller-manager" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825597 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="36865289-7421-41dd-8434-fb2b31a3b974" containerName="controller-manager" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825614 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fb3271e-4d33-4ef0-b269-ca69f698fb66" containerName="route-controller-manager" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825622 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825628 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa80450c-ee41-4f6f-813b-ba7556c751c0" containerName="installer" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.825982 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.833422 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.835633 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.835762 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.835803 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.835910 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.835989 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.839048 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-79b745f746-tkkht"] Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.839705 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.844119 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.845173 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.845587 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.845615 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.846817 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.849764 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl"] Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.850319 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.853211 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79b745f746-tkkht"] Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.858496 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.898191 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36865289-7421-41dd-8434-fb2b31a3b974" path="/var/lib/kubelet/pods/36865289-7421-41dd-8434-fb2b31a3b974/volumes" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.899564 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fb3271e-4d33-4ef0-b269-ca69f698fb66" path="/var/lib/kubelet/pods/3fb3271e-4d33-4ef0-b269-ca69f698fb66/volumes" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.906777 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-config\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.906845 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxs6q\" (UniqueName: \"kubernetes.io/projected/e505abe6-6c95-417a-ad77-539a0d298fae-kube-api-access-cxs6q\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.906869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-client-ca\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.906894 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-client-ca\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.906910 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42bgt\" (UniqueName: \"kubernetes.io/projected/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-kube-api-access-42bgt\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.906928 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-config\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.906965 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-proxy-ca-bundles\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.906987 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-serving-cert\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:22 crc kubenswrapper[4735]: I0215 20:22:22.907001 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e505abe6-6c95-417a-ad77-539a0d298fae-serving-cert\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.008555 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-proxy-ca-bundles\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.008602 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-serving-cert\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.008620 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e505abe6-6c95-417a-ad77-539a0d298fae-serving-cert\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.008640 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-config\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.008691 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxs6q\" (UniqueName: \"kubernetes.io/projected/e505abe6-6c95-417a-ad77-539a0d298fae-kube-api-access-cxs6q\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.008713 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-client-ca\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.008746 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-client-ca\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.008771 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42bgt\" (UniqueName: \"kubernetes.io/projected/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-kube-api-access-42bgt\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.008795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-config\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.009515 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-proxy-ca-bundles\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.009836 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-config\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.010555 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-client-ca\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.011070 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-client-ca\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.012085 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-config\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.017663 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-serving-cert\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.019015 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e505abe6-6c95-417a-ad77-539a0d298fae-serving-cert\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.025454 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxs6q\" (UniqueName: \"kubernetes.io/projected/e505abe6-6c95-417a-ad77-539a0d298fae-kube-api-access-cxs6q\") pod \"controller-manager-79b745f746-tkkht\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.029322 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42bgt\" (UniqueName: \"kubernetes.io/projected/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-kube-api-access-42bgt\") pod \"route-controller-manager-66f8bc4db-jwrnl\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.141090 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.155524 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.533906 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79b745f746-tkkht"] Feb 15 20:22:23 crc kubenswrapper[4735]: I0215 20:22:23.681016 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl"] Feb 15 20:22:23 crc kubenswrapper[4735]: W0215 20:22:23.685612 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cb7f43c_e71a_45cb_9b4e_d1af53c41993.slice/crio-ed3217a72522eec86f9db3f21635b94d71b6536bb883d6015c81244b12f74ee0 WatchSource:0}: Error finding container ed3217a72522eec86f9db3f21635b94d71b6536bb883d6015c81244b12f74ee0: Status 404 returned error can't find the container with id ed3217a72522eec86f9db3f21635b94d71b6536bb883d6015c81244b12f74ee0 Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.328593 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" event={"ID":"e505abe6-6c95-417a-ad77-539a0d298fae","Type":"ContainerStarted","Data":"3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c"} Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.328957 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.328972 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" event={"ID":"e505abe6-6c95-417a-ad77-539a0d298fae","Type":"ContainerStarted","Data":"b1ff52a091d34164ce717cf32d48112109a2c9d2406554c72f865b0a0d5c5aaa"} Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.330842 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" event={"ID":"5cb7f43c-e71a-45cb-9b4e-d1af53c41993","Type":"ContainerStarted","Data":"899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d"} Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.330891 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" event={"ID":"5cb7f43c-e71a-45cb-9b4e-d1af53c41993","Type":"ContainerStarted","Data":"ed3217a72522eec86f9db3f21635b94d71b6536bb883d6015c81244b12f74ee0"} Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.331045 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.336254 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.345935 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" podStartSLOduration=3.345920247 podStartE2EDuration="3.345920247s" podCreationTimestamp="2026-02-15 20:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:22:24.344386074 +0000 UTC m=+352.210401697" watchObservedRunningTime="2026-02-15 20:22:24.345920247 +0000 UTC m=+352.211935870" Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.419765 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.437777 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" podStartSLOduration=3.437759981 podStartE2EDuration="3.437759981s" podCreationTimestamp="2026-02-15 20:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:22:24.399459204 +0000 UTC m=+352.265474827" watchObservedRunningTime="2026-02-15 20:22:24.437759981 +0000 UTC m=+352.303775604" Feb 15 20:22:24 crc kubenswrapper[4735]: I0215 20:22:24.996443 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-79b745f746-tkkht"] Feb 15 20:22:25 crc kubenswrapper[4735]: I0215 20:22:25.007690 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl"] Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.359029 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" podUID="5cb7f43c-e71a-45cb-9b4e-d1af53c41993" containerName="route-controller-manager" containerID="cri-o://899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d" gracePeriod=30 Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.359283 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" podUID="e505abe6-6c95-417a-ad77-539a0d298fae" containerName="controller-manager" containerID="cri-o://3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c" gracePeriod=30 Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.768240 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.811341 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp"] Feb 15 20:22:26 crc kubenswrapper[4735]: E0215 20:22:26.811604 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb7f43c-e71a-45cb-9b4e-d1af53c41993" containerName="route-controller-manager" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.811617 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb7f43c-e71a-45cb-9b4e-d1af53c41993" containerName="route-controller-manager" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.811705 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb7f43c-e71a-45cb-9b4e-d1af53c41993" containerName="route-controller-manager" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.812217 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.840322 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.857808 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-client-ca\") pod \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.857885 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-serving-cert\") pod \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.857929 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-config\") pod \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.858006 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42bgt\" (UniqueName: \"kubernetes.io/projected/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-kube-api-access-42bgt\") pod \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\" (UID: \"5cb7f43c-e71a-45cb-9b4e-d1af53c41993\") " Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.858242 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-client-ca\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.858298 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2c206a-3205-433d-90cb-7a2b27c43287-serving-cert\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.858334 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kmvl\" (UniqueName: \"kubernetes.io/projected/9a2c206a-3205-433d-90cb-7a2b27c43287-kube-api-access-6kmvl\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.858362 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-config\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.861523 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-client-ca" (OuterVolumeSpecName: "client-ca") pod "5cb7f43c-e71a-45cb-9b4e-d1af53c41993" (UID: "5cb7f43c-e71a-45cb-9b4e-d1af53c41993"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.863474 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-config" (OuterVolumeSpecName: "config") pod "5cb7f43c-e71a-45cb-9b4e-d1af53c41993" (UID: "5cb7f43c-e71a-45cb-9b4e-d1af53c41993"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.867093 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5cb7f43c-e71a-45cb-9b4e-d1af53c41993" (UID: "5cb7f43c-e71a-45cb-9b4e-d1af53c41993"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.880013 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-kube-api-access-42bgt" (OuterVolumeSpecName: "kube-api-access-42bgt") pod "5cb7f43c-e71a-45cb-9b4e-d1af53c41993" (UID: "5cb7f43c-e71a-45cb-9b4e-d1af53c41993"). InnerVolumeSpecName "kube-api-access-42bgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.895538 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp"] Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.958902 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-client-ca\") pod \"e505abe6-6c95-417a-ad77-539a0d298fae\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.958984 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxs6q\" (UniqueName: \"kubernetes.io/projected/e505abe6-6c95-417a-ad77-539a0d298fae-kube-api-access-cxs6q\") pod \"e505abe6-6c95-417a-ad77-539a0d298fae\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959042 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-config\") pod \"e505abe6-6c95-417a-ad77-539a0d298fae\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959116 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-proxy-ca-bundles\") pod \"e505abe6-6c95-417a-ad77-539a0d298fae\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959157 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e505abe6-6c95-417a-ad77-539a0d298fae-serving-cert\") pod \"e505abe6-6c95-417a-ad77-539a0d298fae\" (UID: \"e505abe6-6c95-417a-ad77-539a0d298fae\") " Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959411 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-client-ca\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959453 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2c206a-3205-433d-90cb-7a2b27c43287-serving-cert\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959492 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kmvl\" (UniqueName: \"kubernetes.io/projected/9a2c206a-3205-433d-90cb-7a2b27c43287-kube-api-access-6kmvl\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959509 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-config\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959686 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42bgt\" (UniqueName: \"kubernetes.io/projected/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-kube-api-access-42bgt\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959705 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-client-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959715 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.959725 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cb7f43c-e71a-45cb-9b4e-d1af53c41993-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.960182 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e505abe6-6c95-417a-ad77-539a0d298fae" (UID: "e505abe6-6c95-417a-ad77-539a0d298fae"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.960247 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-config" (OuterVolumeSpecName: "config") pod "e505abe6-6c95-417a-ad77-539a0d298fae" (UID: "e505abe6-6c95-417a-ad77-539a0d298fae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.961399 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-client-ca\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.961724 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-client-ca" (OuterVolumeSpecName: "client-ca") pod "e505abe6-6c95-417a-ad77-539a0d298fae" (UID: "e505abe6-6c95-417a-ad77-539a0d298fae"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.961783 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-config\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.963647 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e505abe6-6c95-417a-ad77-539a0d298fae-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e505abe6-6c95-417a-ad77-539a0d298fae" (UID: "e505abe6-6c95-417a-ad77-539a0d298fae"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.964223 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2c206a-3205-433d-90cb-7a2b27c43287-serving-cert\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.964667 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e505abe6-6c95-417a-ad77-539a0d298fae-kube-api-access-cxs6q" (OuterVolumeSpecName: "kube-api-access-cxs6q") pod "e505abe6-6c95-417a-ad77-539a0d298fae" (UID: "e505abe6-6c95-417a-ad77-539a0d298fae"). InnerVolumeSpecName "kube-api-access-cxs6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:26 crc kubenswrapper[4735]: I0215 20:22:26.978030 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kmvl\" (UniqueName: \"kubernetes.io/projected/9a2c206a-3205-433d-90cb-7a2b27c43287-kube-api-access-6kmvl\") pod \"route-controller-manager-675dc77d8c-wkhsp\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.060346 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.060372 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.060383 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e505abe6-6c95-417a-ad77-539a0d298fae-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.060390 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e505abe6-6c95-417a-ad77-539a0d298fae-client-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.060408 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxs6q\" (UniqueName: \"kubernetes.io/projected/e505abe6-6c95-417a-ad77-539a0d298fae-kube-api-access-cxs6q\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.147737 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.369395 4735 generic.go:334] "Generic (PLEG): container finished" podID="5cb7f43c-e71a-45cb-9b4e-d1af53c41993" containerID="899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d" exitCode=0 Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.369465 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" event={"ID":"5cb7f43c-e71a-45cb-9b4e-d1af53c41993","Type":"ContainerDied","Data":"899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d"} Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.369490 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" event={"ID":"5cb7f43c-e71a-45cb-9b4e-d1af53c41993","Type":"ContainerDied","Data":"ed3217a72522eec86f9db3f21635b94d71b6536bb883d6015c81244b12f74ee0"} Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.369507 4735 scope.go:117] "RemoveContainer" containerID="899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.369602 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.372006 4735 generic.go:334] "Generic (PLEG): container finished" podID="e505abe6-6c95-417a-ad77-539a0d298fae" containerID="3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c" exitCode=0 Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.372029 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" event={"ID":"e505abe6-6c95-417a-ad77-539a0d298fae","Type":"ContainerDied","Data":"3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c"} Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.372046 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" event={"ID":"e505abe6-6c95-417a-ad77-539a0d298fae","Type":"ContainerDied","Data":"b1ff52a091d34164ce717cf32d48112109a2c9d2406554c72f865b0a0d5c5aaa"} Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.372082 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79b745f746-tkkht" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.390435 4735 scope.go:117] "RemoveContainer" containerID="899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d" Feb 15 20:22:27 crc kubenswrapper[4735]: E0215 20:22:27.390991 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d\": container with ID starting with 899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d not found: ID does not exist" containerID="899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.391038 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d"} err="failed to get container status \"899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d\": rpc error: code = NotFound desc = could not find container \"899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d\": container with ID starting with 899d5c7a0abb7b516b830669227d4853b3fd4d61ef13e1718e2caa5f3d1c721d not found: ID does not exist" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.391071 4735 scope.go:117] "RemoveContainer" containerID="3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.392210 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl"] Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.394689 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66f8bc4db-jwrnl"] Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.401866 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-79b745f746-tkkht"] Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.404677 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-79b745f746-tkkht"] Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.408042 4735 scope.go:117] "RemoveContainer" containerID="3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c" Feb 15 20:22:27 crc kubenswrapper[4735]: E0215 20:22:27.408614 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c\": container with ID starting with 3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c not found: ID does not exist" containerID="3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.408660 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c"} err="failed to get container status \"3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c\": rpc error: code = NotFound desc = could not find container \"3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c\": container with ID starting with 3d71ce08916ad952b2ff7fbc7483dc906f31a9ff742c23f4cd08cd2b5a8b9d9c not found: ID does not exist" Feb 15 20:22:27 crc kubenswrapper[4735]: I0215 20:22:27.546871 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp"] Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.379293 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" event={"ID":"9a2c206a-3205-433d-90cb-7a2b27c43287","Type":"ContainerStarted","Data":"624d79af41d38c805c1b77d5a9b52a1624fabe80c97935b7bf4981a04e042078"} Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.379591 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" event={"ID":"9a2c206a-3205-433d-90cb-7a2b27c43287","Type":"ContainerStarted","Data":"2ed0455ece6076a398985287df4bdf832c02434d4c245a4aadbfd46c72d981aa"} Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.379608 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.383638 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.423037 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" podStartSLOduration=3.423020351 podStartE2EDuration="3.423020351s" podCreationTimestamp="2026-02-15 20:22:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:22:28.401048443 +0000 UTC m=+356.267064066" watchObservedRunningTime="2026-02-15 20:22:28.423020351 +0000 UTC m=+356.289035974" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.833544 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d9fdddc5-shf2h"] Feb 15 20:22:28 crc kubenswrapper[4735]: E0215 20:22:28.833731 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e505abe6-6c95-417a-ad77-539a0d298fae" containerName="controller-manager" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.833741 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e505abe6-6c95-417a-ad77-539a0d298fae" containerName="controller-manager" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.833839 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e505abe6-6c95-417a-ad77-539a0d298fae" containerName="controller-manager" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.834200 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.836104 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.836289 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.836771 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.837124 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.837344 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.837754 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.843905 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.852370 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d9fdddc5-shf2h"] Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.881807 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-proxy-ca-bundles\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.881866 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4fbt\" (UniqueName: \"kubernetes.io/projected/354e258d-18b9-4ba9-b573-f9f12a484e4a-kube-api-access-g4fbt\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.881938 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-client-ca\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.882019 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-config\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.882066 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/354e258d-18b9-4ba9-b573-f9f12a484e4a-serving-cert\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.894928 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cb7f43c-e71a-45cb-9b4e-d1af53c41993" path="/var/lib/kubelet/pods/5cb7f43c-e71a-45cb-9b4e-d1af53c41993/volumes" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.895739 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e505abe6-6c95-417a-ad77-539a0d298fae" path="/var/lib/kubelet/pods/e505abe6-6c95-417a-ad77-539a0d298fae/volumes" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.983546 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/354e258d-18b9-4ba9-b573-f9f12a484e4a-serving-cert\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.983625 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-proxy-ca-bundles\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.983651 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4fbt\" (UniqueName: \"kubernetes.io/projected/354e258d-18b9-4ba9-b573-f9f12a484e4a-kube-api-access-g4fbt\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.983703 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-client-ca\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.984131 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-config\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.984845 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-client-ca\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.985212 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-config\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.985489 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-proxy-ca-bundles\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:28 crc kubenswrapper[4735]: I0215 20:22:28.990065 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/354e258d-18b9-4ba9-b573-f9f12a484e4a-serving-cert\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:29 crc kubenswrapper[4735]: I0215 20:22:29.008825 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4fbt\" (UniqueName: \"kubernetes.io/projected/354e258d-18b9-4ba9-b573-f9f12a484e4a-kube-api-access-g4fbt\") pod \"controller-manager-6d9fdddc5-shf2h\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:29 crc kubenswrapper[4735]: I0215 20:22:29.147389 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:29 crc kubenswrapper[4735]: I0215 20:22:29.407498 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d9fdddc5-shf2h"] Feb 15 20:22:30 crc kubenswrapper[4735]: I0215 20:22:30.389796 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" event={"ID":"354e258d-18b9-4ba9-b573-f9f12a484e4a","Type":"ContainerStarted","Data":"28d789518da91dfb7484d1a56f9c601888df56f9a04b238320c34c3677673f7d"} Feb 15 20:22:30 crc kubenswrapper[4735]: I0215 20:22:30.390093 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" event={"ID":"354e258d-18b9-4ba9-b573-f9f12a484e4a","Type":"ContainerStarted","Data":"175e5d2dc38abc9c0a16fdcf083d66f5cb4fc1a8711aeabbc9f675e882883597"} Feb 15 20:22:30 crc kubenswrapper[4735]: I0215 20:22:30.407989 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" podStartSLOduration=5.407920364 podStartE2EDuration="5.407920364s" podCreationTimestamp="2026-02-15 20:22:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:22:30.406562975 +0000 UTC m=+358.272578598" watchObservedRunningTime="2026-02-15 20:22:30.407920364 +0000 UTC m=+358.273935987" Feb 15 20:22:31 crc kubenswrapper[4735]: I0215 20:22:31.394790 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:31 crc kubenswrapper[4735]: I0215 20:22:31.399445 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.289540 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp"] Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.290382 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" podUID="9a2c206a-3205-433d-90cb-7a2b27c43287" containerName="route-controller-manager" containerID="cri-o://624d79af41d38c805c1b77d5a9b52a1624fabe80c97935b7bf4981a04e042078" gracePeriod=30 Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.503866 4735 generic.go:334] "Generic (PLEG): container finished" podID="9a2c206a-3205-433d-90cb-7a2b27c43287" containerID="624d79af41d38c805c1b77d5a9b52a1624fabe80c97935b7bf4981a04e042078" exitCode=0 Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.503914 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" event={"ID":"9a2c206a-3205-433d-90cb-7a2b27c43287","Type":"ContainerDied","Data":"624d79af41d38c805c1b77d5a9b52a1624fabe80c97935b7bf4981a04e042078"} Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.717437 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.741126 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-client-ca\") pod \"9a2c206a-3205-433d-90cb-7a2b27c43287\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.741205 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-config\") pod \"9a2c206a-3205-433d-90cb-7a2b27c43287\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.741287 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2c206a-3205-433d-90cb-7a2b27c43287-serving-cert\") pod \"9a2c206a-3205-433d-90cb-7a2b27c43287\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.741326 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kmvl\" (UniqueName: \"kubernetes.io/projected/9a2c206a-3205-433d-90cb-7a2b27c43287-kube-api-access-6kmvl\") pod \"9a2c206a-3205-433d-90cb-7a2b27c43287\" (UID: \"9a2c206a-3205-433d-90cb-7a2b27c43287\") " Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.742341 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-client-ca" (OuterVolumeSpecName: "client-ca") pod "9a2c206a-3205-433d-90cb-7a2b27c43287" (UID: "9a2c206a-3205-433d-90cb-7a2b27c43287"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.743031 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-config" (OuterVolumeSpecName: "config") pod "9a2c206a-3205-433d-90cb-7a2b27c43287" (UID: "9a2c206a-3205-433d-90cb-7a2b27c43287"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.754749 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a2c206a-3205-433d-90cb-7a2b27c43287-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9a2c206a-3205-433d-90cb-7a2b27c43287" (UID: "9a2c206a-3205-433d-90cb-7a2b27c43287"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.754795 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a2c206a-3205-433d-90cb-7a2b27c43287-kube-api-access-6kmvl" (OuterVolumeSpecName: "kube-api-access-6kmvl") pod "9a2c206a-3205-433d-90cb-7a2b27c43287" (UID: "9a2c206a-3205-433d-90cb-7a2b27c43287"). InnerVolumeSpecName "kube-api-access-6kmvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.843052 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2c206a-3205-433d-90cb-7a2b27c43287-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.843338 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kmvl\" (UniqueName: \"kubernetes.io/projected/9a2c206a-3205-433d-90cb-7a2b27c43287-kube-api-access-6kmvl\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.843351 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-client-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:41 crc kubenswrapper[4735]: I0215 20:22:41.843363 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c206a-3205-433d-90cb-7a2b27c43287-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.511316 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" event={"ID":"9a2c206a-3205-433d-90cb-7a2b27c43287","Type":"ContainerDied","Data":"2ed0455ece6076a398985287df4bdf832c02434d4c245a4aadbfd46c72d981aa"} Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.511364 4735 scope.go:117] "RemoveContainer" containerID="624d79af41d38c805c1b77d5a9b52a1624fabe80c97935b7bf4981a04e042078" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.511473 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.551195 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp"] Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.554749 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-675dc77d8c-wkhsp"] Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.849238 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn"] Feb 15 20:22:42 crc kubenswrapper[4735]: E0215 20:22:42.849560 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a2c206a-3205-433d-90cb-7a2b27c43287" containerName="route-controller-manager" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.849587 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a2c206a-3205-433d-90cb-7a2b27c43287" containerName="route-controller-manager" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.849794 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a2c206a-3205-433d-90cb-7a2b27c43287" containerName="route-controller-manager" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.850422 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.853632 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.853826 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.853837 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.853968 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.854088 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.860740 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.865193 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn"] Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.894985 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a2c206a-3205-433d-90cb-7a2b27c43287" path="/var/lib/kubelet/pods/9a2c206a-3205-433d-90cb-7a2b27c43287/volumes" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.955186 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bde6d8e-261a-4372-b394-b996f9856ce3-client-ca\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.955261 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bde6d8e-261a-4372-b394-b996f9856ce3-config\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.955281 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4cwg\" (UniqueName: \"kubernetes.io/projected/3bde6d8e-261a-4372-b394-b996f9856ce3-kube-api-access-l4cwg\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:42 crc kubenswrapper[4735]: I0215 20:22:42.955313 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bde6d8e-261a-4372-b394-b996f9856ce3-serving-cert\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.056364 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bde6d8e-261a-4372-b394-b996f9856ce3-serving-cert\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.056406 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bde6d8e-261a-4372-b394-b996f9856ce3-client-ca\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.056459 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bde6d8e-261a-4372-b394-b996f9856ce3-config\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.056475 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4cwg\" (UniqueName: \"kubernetes.io/projected/3bde6d8e-261a-4372-b394-b996f9856ce3-kube-api-access-l4cwg\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.057738 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bde6d8e-261a-4372-b394-b996f9856ce3-client-ca\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.060007 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bde6d8e-261a-4372-b394-b996f9856ce3-serving-cert\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.060555 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bde6d8e-261a-4372-b394-b996f9856ce3-config\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.083364 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4cwg\" (UniqueName: \"kubernetes.io/projected/3bde6d8e-261a-4372-b394-b996f9856ce3-kube-api-access-l4cwg\") pod \"route-controller-manager-78ddb7876b-xl7bn\" (UID: \"3bde6d8e-261a-4372-b394-b996f9856ce3\") " pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.167881 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:43 crc kubenswrapper[4735]: I0215 20:22:43.578985 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn"] Feb 15 20:22:44 crc kubenswrapper[4735]: I0215 20:22:44.522163 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" event={"ID":"3bde6d8e-261a-4372-b394-b996f9856ce3","Type":"ContainerStarted","Data":"71d51a0ba5d27819932d1f5cfb5f557773a63bc7c001a257cc53eece8688a941"} Feb 15 20:22:44 crc kubenswrapper[4735]: I0215 20:22:44.522505 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" event={"ID":"3bde6d8e-261a-4372-b394-b996f9856ce3","Type":"ContainerStarted","Data":"512a56cd12290f252d0c1dd024dcdcb9abb9810d8b9fed220fc3a3fe0e64613d"} Feb 15 20:22:44 crc kubenswrapper[4735]: I0215 20:22:44.522529 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:44 crc kubenswrapper[4735]: I0215 20:22:44.531915 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" Feb 15 20:22:44 crc kubenswrapper[4735]: I0215 20:22:44.544033 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-78ddb7876b-xl7bn" podStartSLOduration=3.544012086 podStartE2EDuration="3.544012086s" podCreationTimestamp="2026-02-15 20:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:22:44.539376735 +0000 UTC m=+372.405392358" watchObservedRunningTime="2026-02-15 20:22:44.544012086 +0000 UTC m=+372.410027709" Feb 15 20:22:49 crc kubenswrapper[4735]: I0215 20:22:49.680519 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:22:49 crc kubenswrapper[4735]: I0215 20:22:49.681123 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.348303 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4zvg5"] Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.349151 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.388878 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4zvg5"] Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.453886 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fbe07312-58ca-4d5b-8393-a1064b2b23a8-registry-tls\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.454004 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fbe07312-58ca-4d5b-8393-a1064b2b23a8-trusted-ca\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.454022 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fbe07312-58ca-4d5b-8393-a1064b2b23a8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.454042 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fbe07312-58ca-4d5b-8393-a1064b2b23a8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.454082 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fbe07312-58ca-4d5b-8393-a1064b2b23a8-registry-certificates\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.454108 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fbe07312-58ca-4d5b-8393-a1064b2b23a8-bound-sa-token\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.454123 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrvf5\" (UniqueName: \"kubernetes.io/projected/fbe07312-58ca-4d5b-8393-a1064b2b23a8-kube-api-access-xrvf5\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.454148 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.479076 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.554874 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fbe07312-58ca-4d5b-8393-a1064b2b23a8-registry-tls\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.554917 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fbe07312-58ca-4d5b-8393-a1064b2b23a8-trusted-ca\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.554934 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fbe07312-58ca-4d5b-8393-a1064b2b23a8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.554975 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fbe07312-58ca-4d5b-8393-a1064b2b23a8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.555026 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fbe07312-58ca-4d5b-8393-a1064b2b23a8-registry-certificates\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.555060 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrvf5\" (UniqueName: \"kubernetes.io/projected/fbe07312-58ca-4d5b-8393-a1064b2b23a8-kube-api-access-xrvf5\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.555084 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fbe07312-58ca-4d5b-8393-a1064b2b23a8-bound-sa-token\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.555766 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fbe07312-58ca-4d5b-8393-a1064b2b23a8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.556489 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fbe07312-58ca-4d5b-8393-a1064b2b23a8-trusted-ca\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.556822 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fbe07312-58ca-4d5b-8393-a1064b2b23a8-registry-certificates\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.561061 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fbe07312-58ca-4d5b-8393-a1064b2b23a8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.561121 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fbe07312-58ca-4d5b-8393-a1064b2b23a8-registry-tls\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.576057 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fbe07312-58ca-4d5b-8393-a1064b2b23a8-bound-sa-token\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.579129 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrvf5\" (UniqueName: \"kubernetes.io/projected/fbe07312-58ca-4d5b-8393-a1064b2b23a8-kube-api-access-xrvf5\") pod \"image-registry-66df7c8f76-4zvg5\" (UID: \"fbe07312-58ca-4d5b-8393-a1064b2b23a8\") " pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:50 crc kubenswrapper[4735]: I0215 20:22:50.687755 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:51 crc kubenswrapper[4735]: I0215 20:22:51.108617 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4zvg5"] Feb 15 20:22:51 crc kubenswrapper[4735]: I0215 20:22:51.560027 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" event={"ID":"fbe07312-58ca-4d5b-8393-a1064b2b23a8","Type":"ContainerStarted","Data":"42a8ae01ff3e5cfe0aada737c8d40a693b2008831fe7366a444fbc5e48fb74f5"} Feb 15 20:22:51 crc kubenswrapper[4735]: I0215 20:22:51.560072 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" event={"ID":"fbe07312-58ca-4d5b-8393-a1064b2b23a8","Type":"ContainerStarted","Data":"cb0e0cb684e999127cf88fe51a077002615722a08b6858f554d2057328875288"} Feb 15 20:22:51 crc kubenswrapper[4735]: I0215 20:22:51.560205 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:22:51 crc kubenswrapper[4735]: I0215 20:22:51.582013 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" podStartSLOduration=1.581993095 podStartE2EDuration="1.581993095s" podCreationTimestamp="2026-02-15 20:22:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:22:51.581764799 +0000 UTC m=+379.447780442" watchObservedRunningTime="2026-02-15 20:22:51.581993095 +0000 UTC m=+379.448008748" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.520897 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2cm8"] Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.522920 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z2cm8" podUID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerName="registry-server" containerID="cri-o://e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99" gracePeriod=30 Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.533522 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lzl2l"] Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.533725 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lzl2l" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerName="registry-server" containerID="cri-o://ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112" gracePeriod=30 Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.560262 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-527r8"] Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.560446 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" containerID="cri-o://85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261" gracePeriod=30 Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.571682 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hm2zs"] Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.575131 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hm2zs" podUID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerName="registry-server" containerID="cri-o://8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b" gracePeriod=30 Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.576527 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ff4s6"] Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.577402 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.580698 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7vk2f"] Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.580909 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7vk2f" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerName="registry-server" containerID="cri-o://1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82" gracePeriod=30 Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.607148 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ff4s6"] Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.711256 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02a0669b-e47e-4f62-851c-a776d1f43b12-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ff4s6\" (UID: \"02a0669b-e47e-4f62-851c-a776d1f43b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.711625 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a0669b-e47e-4f62-851c-a776d1f43b12-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ff4s6\" (UID: \"02a0669b-e47e-4f62-851c-a776d1f43b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.711674 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55qx6\" (UniqueName: \"kubernetes.io/projected/02a0669b-e47e-4f62-851c-a776d1f43b12-kube-api-access-55qx6\") pod \"marketplace-operator-79b997595-ff4s6\" (UID: \"02a0669b-e47e-4f62-851c-a776d1f43b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.812055 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55qx6\" (UniqueName: \"kubernetes.io/projected/02a0669b-e47e-4f62-851c-a776d1f43b12-kube-api-access-55qx6\") pod \"marketplace-operator-79b997595-ff4s6\" (UID: \"02a0669b-e47e-4f62-851c-a776d1f43b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.812106 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02a0669b-e47e-4f62-851c-a776d1f43b12-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ff4s6\" (UID: \"02a0669b-e47e-4f62-851c-a776d1f43b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.812181 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a0669b-e47e-4f62-851c-a776d1f43b12-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ff4s6\" (UID: \"02a0669b-e47e-4f62-851c-a776d1f43b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.813450 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a0669b-e47e-4f62-851c-a776d1f43b12-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ff4s6\" (UID: \"02a0669b-e47e-4f62-851c-a776d1f43b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.819400 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02a0669b-e47e-4f62-851c-a776d1f43b12-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ff4s6\" (UID: \"02a0669b-e47e-4f62-851c-a776d1f43b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.831586 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55qx6\" (UniqueName: \"kubernetes.io/projected/02a0669b-e47e-4f62-851c-a776d1f43b12-kube-api-access-55qx6\") pod \"marketplace-operator-79b997595-ff4s6\" (UID: \"02a0669b-e47e-4f62-851c-a776d1f43b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:58 crc kubenswrapper[4735]: I0215 20:22:58.902600 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.051441 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.216399 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-utilities\") pod \"31a015ae-51a2-488b-8bfd-4829857d6c7d\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.216493 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-catalog-content\") pod \"31a015ae-51a2-488b-8bfd-4829857d6c7d\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.216524 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsvrk\" (UniqueName: \"kubernetes.io/projected/31a015ae-51a2-488b-8bfd-4829857d6c7d-kube-api-access-wsvrk\") pod \"31a015ae-51a2-488b-8bfd-4829857d6c7d\" (UID: \"31a015ae-51a2-488b-8bfd-4829857d6c7d\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.225798 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31a015ae-51a2-488b-8bfd-4829857d6c7d-kube-api-access-wsvrk" (OuterVolumeSpecName: "kube-api-access-wsvrk") pod "31a015ae-51a2-488b-8bfd-4829857d6c7d" (UID: "31a015ae-51a2-488b-8bfd-4829857d6c7d"). InnerVolumeSpecName "kube-api-access-wsvrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.226397 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-utilities" (OuterVolumeSpecName: "utilities") pod "31a015ae-51a2-488b-8bfd-4829857d6c7d" (UID: "31a015ae-51a2-488b-8bfd-4829857d6c7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.268078 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31a015ae-51a2-488b-8bfd-4829857d6c7d" (UID: "31a015ae-51a2-488b-8bfd-4829857d6c7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.319024 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.319058 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31a015ae-51a2-488b-8bfd-4829857d6c7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.319070 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsvrk\" (UniqueName: \"kubernetes.io/projected/31a015ae-51a2-488b-8bfd-4829857d6c7d-kube-api-access-wsvrk\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.319573 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.324609 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.350766 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.373922 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.420594 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxxxw\" (UniqueName: \"kubernetes.io/projected/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-kube-api-access-mxxxw\") pod \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.420647 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-catalog-content\") pod \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.420785 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-utilities\") pod \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\" (UID: \"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.421150 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-utilities\") pod \"309558c5-0acb-4b54-aedf-bde3c482dec9\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.421206 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-trusted-ca\") pod \"1626b474-7e91-4cde-bab7-a4db753478dd\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.421262 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-utilities\") pod \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.421277 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-catalog-content\") pod \"309558c5-0acb-4b54-aedf-bde3c482dec9\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.422114 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-utilities" (OuterVolumeSpecName: "utilities") pod "309558c5-0acb-4b54-aedf-bde3c482dec9" (UID: "309558c5-0acb-4b54-aedf-bde3c482dec9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.422124 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-utilities" (OuterVolumeSpecName: "utilities") pod "59b9213c-60ff-4ecd-99d9-4c1fe670592c" (UID: "59b9213c-60ff-4ecd-99d9-4c1fe670592c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.422170 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "1626b474-7e91-4cde-bab7-a4db753478dd" (UID: "1626b474-7e91-4cde-bab7-a4db753478dd"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.422241 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tgrt\" (UniqueName: \"kubernetes.io/projected/59b9213c-60ff-4ecd-99d9-4c1fe670592c-kube-api-access-2tgrt\") pod \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.422813 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdrr6\" (UniqueName: \"kubernetes.io/projected/309558c5-0acb-4b54-aedf-bde3c482dec9-kube-api-access-vdrr6\") pod \"309558c5-0acb-4b54-aedf-bde3c482dec9\" (UID: \"309558c5-0acb-4b54-aedf-bde3c482dec9\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.423134 4735 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.423148 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.423156 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.423333 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-utilities" (OuterVolumeSpecName: "utilities") pod "f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" (UID: "f54d62ae-bc10-4f8d-9dd2-29ca888f3a23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.425835 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/309558c5-0acb-4b54-aedf-bde3c482dec9-kube-api-access-vdrr6" (OuterVolumeSpecName: "kube-api-access-vdrr6") pod "309558c5-0acb-4b54-aedf-bde3c482dec9" (UID: "309558c5-0acb-4b54-aedf-bde3c482dec9"). InnerVolumeSpecName "kube-api-access-vdrr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.453425 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-kube-api-access-mxxxw" (OuterVolumeSpecName: "kube-api-access-mxxxw") pod "f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" (UID: "f54d62ae-bc10-4f8d-9dd2-29ca888f3a23"). InnerVolumeSpecName "kube-api-access-mxxxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.455034 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b9213c-60ff-4ecd-99d9-4c1fe670592c-kube-api-access-2tgrt" (OuterVolumeSpecName: "kube-api-access-2tgrt") pod "59b9213c-60ff-4ecd-99d9-4c1fe670592c" (UID: "59b9213c-60ff-4ecd-99d9-4c1fe670592c"). InnerVolumeSpecName "kube-api-access-2tgrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.476469 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" (UID: "f54d62ae-bc10-4f8d-9dd2-29ca888f3a23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.524061 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-catalog-content\") pod \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\" (UID: \"59b9213c-60ff-4ecd-99d9-4c1fe670592c\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.524128 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-operator-metrics\") pod \"1626b474-7e91-4cde-bab7-a4db753478dd\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.524229 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dfcj\" (UniqueName: \"kubernetes.io/projected/1626b474-7e91-4cde-bab7-a4db753478dd-kube-api-access-9dfcj\") pod \"1626b474-7e91-4cde-bab7-a4db753478dd\" (UID: \"1626b474-7e91-4cde-bab7-a4db753478dd\") " Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.524365 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdrr6\" (UniqueName: \"kubernetes.io/projected/309558c5-0acb-4b54-aedf-bde3c482dec9-kube-api-access-vdrr6\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.524377 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxxxw\" (UniqueName: \"kubernetes.io/projected/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-kube-api-access-mxxxw\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.524387 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.524396 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.524405 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tgrt\" (UniqueName: \"kubernetes.io/projected/59b9213c-60ff-4ecd-99d9-4c1fe670592c-kube-api-access-2tgrt\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.528310 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1626b474-7e91-4cde-bab7-a4db753478dd-kube-api-access-9dfcj" (OuterVolumeSpecName: "kube-api-access-9dfcj") pod "1626b474-7e91-4cde-bab7-a4db753478dd" (UID: "1626b474-7e91-4cde-bab7-a4db753478dd"). InnerVolumeSpecName "kube-api-access-9dfcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.529923 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "1626b474-7e91-4cde-bab7-a4db753478dd" (UID: "1626b474-7e91-4cde-bab7-a4db753478dd"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.545139 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ff4s6"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.569019 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59b9213c-60ff-4ecd-99d9-4c1fe670592c" (UID: "59b9213c-60ff-4ecd-99d9-4c1fe670592c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.589919 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "309558c5-0acb-4b54-aedf-bde3c482dec9" (UID: "309558c5-0acb-4b54-aedf-bde3c482dec9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.619943 4735 generic.go:334] "Generic (PLEG): container finished" podID="1626b474-7e91-4cde-bab7-a4db753478dd" containerID="85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261" exitCode=0 Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.620019 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" event={"ID":"1626b474-7e91-4cde-bab7-a4db753478dd","Type":"ContainerDied","Data":"85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.620046 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" event={"ID":"1626b474-7e91-4cde-bab7-a4db753478dd","Type":"ContainerDied","Data":"331dd034b95df0dea652eec75c06b10d23cb27166085827d1292ea96443e3366"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.620053 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-527r8" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.620061 4735 scope.go:117] "RemoveContainer" containerID="85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.625992 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dfcj\" (UniqueName: \"kubernetes.io/projected/1626b474-7e91-4cde-bab7-a4db753478dd-kube-api-access-9dfcj\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.626018 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b9213c-60ff-4ecd-99d9-4c1fe670592c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.626029 4735 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1626b474-7e91-4cde-bab7-a4db753478dd-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.626039 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/309558c5-0acb-4b54-aedf-bde3c482dec9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.628845 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" event={"ID":"02a0669b-e47e-4f62-851c-a776d1f43b12","Type":"ContainerStarted","Data":"c4015777fe63b5b8372ed43b036f98cd5d6454944980355b58ec71df7b8fdde4"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.633047 4735 generic.go:334] "Generic (PLEG): container finished" podID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerID="8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b" exitCode=0 Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.633088 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm2zs" event={"ID":"59b9213c-60ff-4ecd-99d9-4c1fe670592c","Type":"ContainerDied","Data":"8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.633106 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm2zs" event={"ID":"59b9213c-60ff-4ecd-99d9-4c1fe670592c","Type":"ContainerDied","Data":"e1dc20bdcf999f7f3033724a0132e6c49704fa9dc3dbf2a881e2bb1186fc2a8e"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.633159 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hm2zs" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.643410 4735 generic.go:334] "Generic (PLEG): container finished" podID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerID="ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112" exitCode=0 Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.643550 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzl2l" event={"ID":"31a015ae-51a2-488b-8bfd-4829857d6c7d","Type":"ContainerDied","Data":"ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.643624 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzl2l" event={"ID":"31a015ae-51a2-488b-8bfd-4829857d6c7d","Type":"ContainerDied","Data":"14a6772ae1869b63804bdcd96fb8e8336faf4f28f68af2c0cc165f7e405b4936"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.643528 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzl2l" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.646606 4735 scope.go:117] "RemoveContainer" containerID="be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.664731 4735 generic.go:334] "Generic (PLEG): container finished" podID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerID="e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99" exitCode=0 Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.664836 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2cm8" event={"ID":"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23","Type":"ContainerDied","Data":"e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.664868 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2cm8" event={"ID":"f54d62ae-bc10-4f8d-9dd2-29ca888f3a23","Type":"ContainerDied","Data":"84e25a0254db6ef7dee717854d2c2820780f0346f7062c33e70662acce2a3c1a"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.665037 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2cm8" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.669751 4735 generic.go:334] "Generic (PLEG): container finished" podID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerID="1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82" exitCode=0 Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.669799 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vk2f" event={"ID":"309558c5-0acb-4b54-aedf-bde3c482dec9","Type":"ContainerDied","Data":"1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.669823 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vk2f" event={"ID":"309558c5-0acb-4b54-aedf-bde3c482dec9","Type":"ContainerDied","Data":"5dc97d1d176104400775f5bbfb7f9e3cda6b2980f294bfa976576c195737258a"} Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.669891 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vk2f" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.676710 4735 scope.go:117] "RemoveContainer" containerID="85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.677708 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261\": container with ID starting with 85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261 not found: ID does not exist" containerID="85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.677737 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261"} err="failed to get container status \"85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261\": rpc error: code = NotFound desc = could not find container \"85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261\": container with ID starting with 85890c4e2b22a41463bc086bc704c18b7fd2b251add1fa63ddde83dbfa3b6261 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.677756 4735 scope.go:117] "RemoveContainer" containerID="be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.678068 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835\": container with ID starting with be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835 not found: ID does not exist" containerID="be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.678114 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835"} err="failed to get container status \"be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835\": rpc error: code = NotFound desc = could not find container \"be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835\": container with ID starting with be3854bd4fdeee8c35a6b414364d02d3c2331cde42fe077e5f282fd5d834d835 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.678127 4735 scope.go:117] "RemoveContainer" containerID="8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.680492 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-527r8"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.684061 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-527r8"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.746641 4735 scope.go:117] "RemoveContainer" containerID="bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.771248 4735 scope.go:117] "RemoveContainer" containerID="32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.774521 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lzl2l"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.790022 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lzl2l"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.799018 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7vk2f"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.799214 4735 scope.go:117] "RemoveContainer" containerID="8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.801359 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b\": container with ID starting with 8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b not found: ID does not exist" containerID="8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.801409 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b"} err="failed to get container status \"8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b\": rpc error: code = NotFound desc = could not find container \"8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b\": container with ID starting with 8a1ed7cb0f8bed514ca7fbc0e54c0f663421d7c8e0aa1ce9c8e200db7e4b7a1b not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.801438 4735 scope.go:117] "RemoveContainer" containerID="bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.801892 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600\": container with ID starting with bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600 not found: ID does not exist" containerID="bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.801960 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600"} err="failed to get container status \"bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600\": rpc error: code = NotFound desc = could not find container \"bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600\": container with ID starting with bf0556799c7fd011a7702f628f02965f46e1b6338fae5b7ab4fb85e473b0d600 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.801992 4735 scope.go:117] "RemoveContainer" containerID="32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.803492 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c\": container with ID starting with 32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c not found: ID does not exist" containerID="32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.803536 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c"} err="failed to get container status \"32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c\": rpc error: code = NotFound desc = could not find container \"32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c\": container with ID starting with 32a86020241583b04ae0d705de74e52a288a8492839f7b81480fabdb09b7828c not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.803557 4735 scope.go:117] "RemoveContainer" containerID="ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.806700 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7vk2f"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.812790 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2cm8"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.816184 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z2cm8"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.821818 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hm2zs"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.825391 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hm2zs"] Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.825607 4735 scope.go:117] "RemoveContainer" containerID="ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.845149 4735 scope.go:117] "RemoveContainer" containerID="9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.857005 4735 scope.go:117] "RemoveContainer" containerID="ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.857364 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112\": container with ID starting with ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112 not found: ID does not exist" containerID="ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.857395 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112"} err="failed to get container status \"ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112\": rpc error: code = NotFound desc = could not find container \"ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112\": container with ID starting with ad2a369c2c4b6eb93bfe83c4b6700b877b6f0acd952cdc7be9d894116ebd8112 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.857436 4735 scope.go:117] "RemoveContainer" containerID="ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.857745 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9\": container with ID starting with ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9 not found: ID does not exist" containerID="ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.857799 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9"} err="failed to get container status \"ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9\": rpc error: code = NotFound desc = could not find container \"ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9\": container with ID starting with ab04a8c4d070ed544d493a243d5fa9c318702aa71fec3a016dfef9779085b9d9 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.857837 4735 scope.go:117] "RemoveContainer" containerID="9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.858572 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba\": container with ID starting with 9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba not found: ID does not exist" containerID="9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.858620 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba"} err="failed to get container status \"9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba\": rpc error: code = NotFound desc = could not find container \"9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba\": container with ID starting with 9f515dedb55b9f4f00bebdbe3aadc5c1cf841bc52995588ea9cff825de5d6bba not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.858663 4735 scope.go:117] "RemoveContainer" containerID="e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.877104 4735 scope.go:117] "RemoveContainer" containerID="47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.890385 4735 scope.go:117] "RemoveContainer" containerID="8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.915625 4735 scope.go:117] "RemoveContainer" containerID="e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.915996 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99\": container with ID starting with e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99 not found: ID does not exist" containerID="e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.916040 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99"} err="failed to get container status \"e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99\": rpc error: code = NotFound desc = could not find container \"e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99\": container with ID starting with e60bb3e977b98c50b2948afb0b6a0440c3ae65613598da9296eb8d6ad610ad99 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.916068 4735 scope.go:117] "RemoveContainer" containerID="47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.916349 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594\": container with ID starting with 47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594 not found: ID does not exist" containerID="47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.916468 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594"} err="failed to get container status \"47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594\": rpc error: code = NotFound desc = could not find container \"47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594\": container with ID starting with 47378e0dfa455abac83489f4ffe1825c09e6b3ca9195b7eaedef96316952d594 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.916577 4735 scope.go:117] "RemoveContainer" containerID="8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.917433 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801\": container with ID starting with 8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801 not found: ID does not exist" containerID="8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.917457 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801"} err="failed to get container status \"8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801\": rpc error: code = NotFound desc = could not find container \"8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801\": container with ID starting with 8883727a5589f85be564c0e07e082c9e253b7817a5d2eb69154a1b1201565801 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.917472 4735 scope.go:117] "RemoveContainer" containerID="1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.936764 4735 scope.go:117] "RemoveContainer" containerID="b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.955935 4735 scope.go:117] "RemoveContainer" containerID="613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.969497 4735 scope.go:117] "RemoveContainer" containerID="1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.970606 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82\": container with ID starting with 1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82 not found: ID does not exist" containerID="1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.970657 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82"} err="failed to get container status \"1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82\": rpc error: code = NotFound desc = could not find container \"1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82\": container with ID starting with 1267b182e49a3519ae16b9b8e1fb9889d6d283a6bb2ddec8f3e4aaa595bc6f82 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.970698 4735 scope.go:117] "RemoveContainer" containerID="b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.971090 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50\": container with ID starting with b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50 not found: ID does not exist" containerID="b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.971128 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50"} err="failed to get container status \"b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50\": rpc error: code = NotFound desc = could not find container \"b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50\": container with ID starting with b35bd95c2abf574e1dda20817186f72e11bc32e69bd64396e763f0a43ff2af50 not found: ID does not exist" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.971149 4735 scope.go:117] "RemoveContainer" containerID="613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26" Feb 15 20:22:59 crc kubenswrapper[4735]: E0215 20:22:59.971565 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26\": container with ID starting with 613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26 not found: ID does not exist" containerID="613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26" Feb 15 20:22:59 crc kubenswrapper[4735]: I0215 20:22:59.971587 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26"} err="failed to get container status \"613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26\": rpc error: code = NotFound desc = could not find container \"613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26\": container with ID starting with 613772d69ea36723ed1a4f526994bcc98d806d2b8676789a80183b9783bb3d26 not found: ID does not exist" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.680444 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" event={"ID":"02a0669b-e47e-4f62-851c-a776d1f43b12","Type":"ContainerStarted","Data":"ff128ed480c2dccb402cac133424ceb8ea45f70bb02c135a74c505bb4e9b61be"} Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.680905 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.685441 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.698663 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ff4s6" podStartSLOduration=2.698650076 podStartE2EDuration="2.698650076s" podCreationTimestamp="2026-02-15 20:22:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:23:00.696907517 +0000 UTC m=+388.562923140" watchObservedRunningTime="2026-02-15 20:23:00.698650076 +0000 UTC m=+388.564665699" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.747756 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d4x4r"] Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748080 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748104 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748117 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerName="extract-content" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748126 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerName="extract-content" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748145 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerName="extract-content" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748154 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerName="extract-content" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748168 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerName="extract-content" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748176 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerName="extract-content" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748190 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748199 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748211 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748222 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748233 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerName="extract-utilities" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748242 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerName="extract-utilities" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748260 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerName="extract-utilities" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748269 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerName="extract-utilities" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748283 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748293 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748312 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerName="extract-utilities" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748323 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerName="extract-utilities" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748333 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748343 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748359 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerName="extract-utilities" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748369 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerName="extract-utilities" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748383 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerName="extract-content" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748394 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerName="extract-content" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748538 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748556 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748568 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748590 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748625 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748640 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" containerName="registry-server" Feb 15 20:23:00 crc kubenswrapper[4735]: E0215 20:23:00.748794 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.748807 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" containerName="marketplace-operator" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.749745 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.750825 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4x4r"] Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.752270 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.842977 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af33352e-df7a-4fb2-af44-d084a141e402-utilities\") pod \"certified-operators-d4x4r\" (UID: \"af33352e-df7a-4fb2-af44-d084a141e402\") " pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.843027 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwppb\" (UniqueName: \"kubernetes.io/projected/af33352e-df7a-4fb2-af44-d084a141e402-kube-api-access-zwppb\") pod \"certified-operators-d4x4r\" (UID: \"af33352e-df7a-4fb2-af44-d084a141e402\") " pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.843075 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af33352e-df7a-4fb2-af44-d084a141e402-catalog-content\") pod \"certified-operators-d4x4r\" (UID: \"af33352e-df7a-4fb2-af44-d084a141e402\") " pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.897784 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1626b474-7e91-4cde-bab7-a4db753478dd" path="/var/lib/kubelet/pods/1626b474-7e91-4cde-bab7-a4db753478dd/volumes" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.898422 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="309558c5-0acb-4b54-aedf-bde3c482dec9" path="/var/lib/kubelet/pods/309558c5-0acb-4b54-aedf-bde3c482dec9/volumes" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.899330 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31a015ae-51a2-488b-8bfd-4829857d6c7d" path="/var/lib/kubelet/pods/31a015ae-51a2-488b-8bfd-4829857d6c7d/volumes" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.900592 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59b9213c-60ff-4ecd-99d9-4c1fe670592c" path="/var/lib/kubelet/pods/59b9213c-60ff-4ecd-99d9-4c1fe670592c/volumes" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.901357 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f54d62ae-bc10-4f8d-9dd2-29ca888f3a23" path="/var/lib/kubelet/pods/f54d62ae-bc10-4f8d-9dd2-29ca888f3a23/volumes" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.938249 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bkthb"] Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.951293 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.952361 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af33352e-df7a-4fb2-af44-d084a141e402-utilities\") pod \"certified-operators-d4x4r\" (UID: \"af33352e-df7a-4fb2-af44-d084a141e402\") " pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.952431 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwppb\" (UniqueName: \"kubernetes.io/projected/af33352e-df7a-4fb2-af44-d084a141e402-kube-api-access-zwppb\") pod \"certified-operators-d4x4r\" (UID: \"af33352e-df7a-4fb2-af44-d084a141e402\") " pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.952470 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af33352e-df7a-4fb2-af44-d084a141e402-catalog-content\") pod \"certified-operators-d4x4r\" (UID: \"af33352e-df7a-4fb2-af44-d084a141e402\") " pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.952938 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af33352e-df7a-4fb2-af44-d084a141e402-catalog-content\") pod \"certified-operators-d4x4r\" (UID: \"af33352e-df7a-4fb2-af44-d084a141e402\") " pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.955427 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af33352e-df7a-4fb2-af44-d084a141e402-utilities\") pod \"certified-operators-d4x4r\" (UID: \"af33352e-df7a-4fb2-af44-d084a141e402\") " pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.960369 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkthb"] Feb 15 20:23:00 crc kubenswrapper[4735]: I0215 20:23:00.996493 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.015895 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwppb\" (UniqueName: \"kubernetes.io/projected/af33352e-df7a-4fb2-af44-d084a141e402-kube-api-access-zwppb\") pod \"certified-operators-d4x4r\" (UID: \"af33352e-df7a-4fb2-af44-d084a141e402\") " pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.054652 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceaf5522-aed5-418f-b483-4c0b945dc615-catalog-content\") pod \"community-operators-bkthb\" (UID: \"ceaf5522-aed5-418f-b483-4c0b945dc615\") " pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.055104 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceaf5522-aed5-418f-b483-4c0b945dc615-utilities\") pod \"community-operators-bkthb\" (UID: \"ceaf5522-aed5-418f-b483-4c0b945dc615\") " pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.055412 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjdr2\" (UniqueName: \"kubernetes.io/projected/ceaf5522-aed5-418f-b483-4c0b945dc615-kube-api-access-mjdr2\") pod \"community-operators-bkthb\" (UID: \"ceaf5522-aed5-418f-b483-4c0b945dc615\") " pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.072350 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.156760 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceaf5522-aed5-418f-b483-4c0b945dc615-utilities\") pod \"community-operators-bkthb\" (UID: \"ceaf5522-aed5-418f-b483-4c0b945dc615\") " pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.156807 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjdr2\" (UniqueName: \"kubernetes.io/projected/ceaf5522-aed5-418f-b483-4c0b945dc615-kube-api-access-mjdr2\") pod \"community-operators-bkthb\" (UID: \"ceaf5522-aed5-418f-b483-4c0b945dc615\") " pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.156842 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceaf5522-aed5-418f-b483-4c0b945dc615-catalog-content\") pod \"community-operators-bkthb\" (UID: \"ceaf5522-aed5-418f-b483-4c0b945dc615\") " pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.157362 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceaf5522-aed5-418f-b483-4c0b945dc615-catalog-content\") pod \"community-operators-bkthb\" (UID: \"ceaf5522-aed5-418f-b483-4c0b945dc615\") " pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.157473 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceaf5522-aed5-418f-b483-4c0b945dc615-utilities\") pod \"community-operators-bkthb\" (UID: \"ceaf5522-aed5-418f-b483-4c0b945dc615\") " pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.174880 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjdr2\" (UniqueName: \"kubernetes.io/projected/ceaf5522-aed5-418f-b483-4c0b945dc615-kube-api-access-mjdr2\") pod \"community-operators-bkthb\" (UID: \"ceaf5522-aed5-418f-b483-4c0b945dc615\") " pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.272742 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d9fdddc5-shf2h"] Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.272962 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" podUID="354e258d-18b9-4ba9-b573-f9f12a484e4a" containerName="controller-manager" containerID="cri-o://28d789518da91dfb7484d1a56f9c601888df56f9a04b238320c34c3677673f7d" gracePeriod=30 Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.315979 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.475301 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4x4r"] Feb 15 20:23:01 crc kubenswrapper[4735]: W0215 20:23:01.494728 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf33352e_df7a_4fb2_af44_d084a141e402.slice/crio-806e374b0a60b42b05e76193fb50cf2247e45ac5e99f26fd622be17ac855f082 WatchSource:0}: Error finding container 806e374b0a60b42b05e76193fb50cf2247e45ac5e99f26fd622be17ac855f082: Status 404 returned error can't find the container with id 806e374b0a60b42b05e76193fb50cf2247e45ac5e99f26fd622be17ac855f082 Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.690723 4735 generic.go:334] "Generic (PLEG): container finished" podID="af33352e-df7a-4fb2-af44-d084a141e402" containerID="7b6ac61431945218169f9b66be4ebb5d0423abddfba5891d4e0cd59541ec84f0" exitCode=0 Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.691085 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4x4r" event={"ID":"af33352e-df7a-4fb2-af44-d084a141e402","Type":"ContainerDied","Data":"7b6ac61431945218169f9b66be4ebb5d0423abddfba5891d4e0cd59541ec84f0"} Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.692162 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4x4r" event={"ID":"af33352e-df7a-4fb2-af44-d084a141e402","Type":"ContainerStarted","Data":"806e374b0a60b42b05e76193fb50cf2247e45ac5e99f26fd622be17ac855f082"} Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.694550 4735 generic.go:334] "Generic (PLEG): container finished" podID="354e258d-18b9-4ba9-b573-f9f12a484e4a" containerID="28d789518da91dfb7484d1a56f9c601888df56f9a04b238320c34c3677673f7d" exitCode=0 Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.695119 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" event={"ID":"354e258d-18b9-4ba9-b573-f9f12a484e4a","Type":"ContainerDied","Data":"28d789518da91dfb7484d1a56f9c601888df56f9a04b238320c34c3677673f7d"} Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.745146 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkthb"] Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.861168 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.866493 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-client-ca\") pod \"354e258d-18b9-4ba9-b573-f9f12a484e4a\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.866530 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4fbt\" (UniqueName: \"kubernetes.io/projected/354e258d-18b9-4ba9-b573-f9f12a484e4a-kube-api-access-g4fbt\") pod \"354e258d-18b9-4ba9-b573-f9f12a484e4a\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.866575 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/354e258d-18b9-4ba9-b573-f9f12a484e4a-serving-cert\") pod \"354e258d-18b9-4ba9-b573-f9f12a484e4a\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.866621 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-config\") pod \"354e258d-18b9-4ba9-b573-f9f12a484e4a\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.866703 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-proxy-ca-bundles\") pod \"354e258d-18b9-4ba9-b573-f9f12a484e4a\" (UID: \"354e258d-18b9-4ba9-b573-f9f12a484e4a\") " Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.867915 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "354e258d-18b9-4ba9-b573-f9f12a484e4a" (UID: "354e258d-18b9-4ba9-b573-f9f12a484e4a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.868366 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-client-ca" (OuterVolumeSpecName: "client-ca") pod "354e258d-18b9-4ba9-b573-f9f12a484e4a" (UID: "354e258d-18b9-4ba9-b573-f9f12a484e4a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.869229 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-config" (OuterVolumeSpecName: "config") pod "354e258d-18b9-4ba9-b573-f9f12a484e4a" (UID: "354e258d-18b9-4ba9-b573-f9f12a484e4a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.874407 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/354e258d-18b9-4ba9-b573-f9f12a484e4a-kube-api-access-g4fbt" (OuterVolumeSpecName: "kube-api-access-g4fbt") pod "354e258d-18b9-4ba9-b573-f9f12a484e4a" (UID: "354e258d-18b9-4ba9-b573-f9f12a484e4a"). InnerVolumeSpecName "kube-api-access-g4fbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.875032 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354e258d-18b9-4ba9-b573-f9f12a484e4a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "354e258d-18b9-4ba9-b573-f9f12a484e4a" (UID: "354e258d-18b9-4ba9-b573-f9f12a484e4a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.967740 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.967772 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.967783 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/354e258d-18b9-4ba9-b573-f9f12a484e4a-client-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.967792 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4fbt\" (UniqueName: \"kubernetes.io/projected/354e258d-18b9-4ba9-b573-f9f12a484e4a-kube-api-access-g4fbt\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:01 crc kubenswrapper[4735]: I0215 20:23:01.967801 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/354e258d-18b9-4ba9-b573-f9f12a484e4a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.700064 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" event={"ID":"354e258d-18b9-4ba9-b573-f9f12a484e4a","Type":"ContainerDied","Data":"175e5d2dc38abc9c0a16fdcf083d66f5cb4fc1a8711aeabbc9f675e882883597"} Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.700422 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d9fdddc5-shf2h" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.705691 4735 scope.go:117] "RemoveContainer" containerID="28d789518da91dfb7484d1a56f9c601888df56f9a04b238320c34c3677673f7d" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.712558 4735 generic.go:334] "Generic (PLEG): container finished" podID="af33352e-df7a-4fb2-af44-d084a141e402" containerID="85f0461d82b0e42b6d786bc18f6b4c3e055a60bb22a4e294952ea3df54ed7dad" exitCode=0 Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.712603 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4x4r" event={"ID":"af33352e-df7a-4fb2-af44-d084a141e402","Type":"ContainerDied","Data":"85f0461d82b0e42b6d786bc18f6b4c3e055a60bb22a4e294952ea3df54ed7dad"} Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.715532 4735 generic.go:334] "Generic (PLEG): container finished" podID="ceaf5522-aed5-418f-b483-4c0b945dc615" containerID="b708ad8d710a66181e20eed345e425402bbbd07a554340f8b6131f725fab9a27" exitCode=0 Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.716440 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkthb" event={"ID":"ceaf5522-aed5-418f-b483-4c0b945dc615","Type":"ContainerDied","Data":"b708ad8d710a66181e20eed345e425402bbbd07a554340f8b6131f725fab9a27"} Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.716458 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkthb" event={"ID":"ceaf5522-aed5-418f-b483-4c0b945dc615","Type":"ContainerStarted","Data":"9920fd328fa227468e36be1ad505aa0a4bebc8b87e187000ea40a4f50fd362fe"} Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.750935 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d9fdddc5-shf2h"] Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.754506 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6d9fdddc5-shf2h"] Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.858142 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7f44b59745-xszbh"] Feb 15 20:23:02 crc kubenswrapper[4735]: E0215 20:23:02.858342 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="354e258d-18b9-4ba9-b573-f9f12a484e4a" containerName="controller-manager" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.858353 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="354e258d-18b9-4ba9-b573-f9f12a484e4a" containerName="controller-manager" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.858444 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="354e258d-18b9-4ba9-b573-f9f12a484e4a" containerName="controller-manager" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.858758 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.861332 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.861355 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.861341 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.861422 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.864829 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.865124 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.871290 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.878469 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67c85582-9e2c-43ed-9a9f-dc46ce546421-client-ca\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.878606 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67c85582-9e2c-43ed-9a9f-dc46ce546421-config\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.878689 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67c85582-9e2c-43ed-9a9f-dc46ce546421-proxy-ca-bundles\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.878721 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67c85582-9e2c-43ed-9a9f-dc46ce546421-serving-cert\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.878802 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48p2p\" (UniqueName: \"kubernetes.io/projected/67c85582-9e2c-43ed-9a9f-dc46ce546421-kube-api-access-48p2p\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.923193 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="354e258d-18b9-4ba9-b573-f9f12a484e4a" path="/var/lib/kubelet/pods/354e258d-18b9-4ba9-b573-f9f12a484e4a/volumes" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.923754 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f44b59745-xszbh"] Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.980419 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67c85582-9e2c-43ed-9a9f-dc46ce546421-serving-cert\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.980497 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48p2p\" (UniqueName: \"kubernetes.io/projected/67c85582-9e2c-43ed-9a9f-dc46ce546421-kube-api-access-48p2p\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.980564 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67c85582-9e2c-43ed-9a9f-dc46ce546421-client-ca\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.980593 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67c85582-9e2c-43ed-9a9f-dc46ce546421-config\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.980643 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67c85582-9e2c-43ed-9a9f-dc46ce546421-proxy-ca-bundles\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.982403 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67c85582-9e2c-43ed-9a9f-dc46ce546421-client-ca\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.983141 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67c85582-9e2c-43ed-9a9f-dc46ce546421-config\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.984348 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67c85582-9e2c-43ed-9a9f-dc46ce546421-proxy-ca-bundles\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.989202 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67c85582-9e2c-43ed-9a9f-dc46ce546421-serving-cert\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:02 crc kubenswrapper[4735]: I0215 20:23:02.995088 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48p2p\" (UniqueName: \"kubernetes.io/projected/67c85582-9e2c-43ed-9a9f-dc46ce546421-kube-api-access-48p2p\") pod \"controller-manager-7f44b59745-xszbh\" (UID: \"67c85582-9e2c-43ed-9a9f-dc46ce546421\") " pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.141037 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gvxkv"] Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.144829 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.150022 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gvxkv"] Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.151382 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.177931 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.284198 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41f0acf-7b74-44f7-92a4-30fb79ccea7a-catalog-content\") pod \"redhat-marketplace-gvxkv\" (UID: \"f41f0acf-7b74-44f7-92a4-30fb79ccea7a\") " pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.284631 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2nhg\" (UniqueName: \"kubernetes.io/projected/f41f0acf-7b74-44f7-92a4-30fb79ccea7a-kube-api-access-z2nhg\") pod \"redhat-marketplace-gvxkv\" (UID: \"f41f0acf-7b74-44f7-92a4-30fb79ccea7a\") " pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.285166 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41f0acf-7b74-44f7-92a4-30fb79ccea7a-utilities\") pod \"redhat-marketplace-gvxkv\" (UID: \"f41f0acf-7b74-44f7-92a4-30fb79ccea7a\") " pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.347320 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c4stk"] Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.348999 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.354983 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.366680 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4stk"] Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.386622 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec1828e-1499-4e18-8b2d-5cfbe9a6d457-catalog-content\") pod \"redhat-operators-c4stk\" (UID: \"aec1828e-1499-4e18-8b2d-5cfbe9a6d457\") " pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.386676 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec1828e-1499-4e18-8b2d-5cfbe9a6d457-utilities\") pod \"redhat-operators-c4stk\" (UID: \"aec1828e-1499-4e18-8b2d-5cfbe9a6d457\") " pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.386703 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41f0acf-7b74-44f7-92a4-30fb79ccea7a-catalog-content\") pod \"redhat-marketplace-gvxkv\" (UID: \"f41f0acf-7b74-44f7-92a4-30fb79ccea7a\") " pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.386734 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vppw7\" (UniqueName: \"kubernetes.io/projected/aec1828e-1499-4e18-8b2d-5cfbe9a6d457-kube-api-access-vppw7\") pod \"redhat-operators-c4stk\" (UID: \"aec1828e-1499-4e18-8b2d-5cfbe9a6d457\") " pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.386763 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2nhg\" (UniqueName: \"kubernetes.io/projected/f41f0acf-7b74-44f7-92a4-30fb79ccea7a-kube-api-access-z2nhg\") pod \"redhat-marketplace-gvxkv\" (UID: \"f41f0acf-7b74-44f7-92a4-30fb79ccea7a\") " pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.386781 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41f0acf-7b74-44f7-92a4-30fb79ccea7a-utilities\") pod \"redhat-marketplace-gvxkv\" (UID: \"f41f0acf-7b74-44f7-92a4-30fb79ccea7a\") " pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.387197 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41f0acf-7b74-44f7-92a4-30fb79ccea7a-utilities\") pod \"redhat-marketplace-gvxkv\" (UID: \"f41f0acf-7b74-44f7-92a4-30fb79ccea7a\") " pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.387431 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41f0acf-7b74-44f7-92a4-30fb79ccea7a-catalog-content\") pod \"redhat-marketplace-gvxkv\" (UID: \"f41f0acf-7b74-44f7-92a4-30fb79ccea7a\") " pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.386363 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f44b59745-xszbh"] Feb 15 20:23:03 crc kubenswrapper[4735]: W0215 20:23:03.406202 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67c85582_9e2c_43ed_9a9f_dc46ce546421.slice/crio-7c0a13885cef77766b831303a553c9a5eac5b8953b189b6193d68ad34afe9d5c WatchSource:0}: Error finding container 7c0a13885cef77766b831303a553c9a5eac5b8953b189b6193d68ad34afe9d5c: Status 404 returned error can't find the container with id 7c0a13885cef77766b831303a553c9a5eac5b8953b189b6193d68ad34afe9d5c Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.408184 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2nhg\" (UniqueName: \"kubernetes.io/projected/f41f0acf-7b74-44f7-92a4-30fb79ccea7a-kube-api-access-z2nhg\") pod \"redhat-marketplace-gvxkv\" (UID: \"f41f0acf-7b74-44f7-92a4-30fb79ccea7a\") " pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.469164 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.487620 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec1828e-1499-4e18-8b2d-5cfbe9a6d457-catalog-content\") pod \"redhat-operators-c4stk\" (UID: \"aec1828e-1499-4e18-8b2d-5cfbe9a6d457\") " pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.487987 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec1828e-1499-4e18-8b2d-5cfbe9a6d457-catalog-content\") pod \"redhat-operators-c4stk\" (UID: \"aec1828e-1499-4e18-8b2d-5cfbe9a6d457\") " pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.488051 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec1828e-1499-4e18-8b2d-5cfbe9a6d457-utilities\") pod \"redhat-operators-c4stk\" (UID: \"aec1828e-1499-4e18-8b2d-5cfbe9a6d457\") " pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.488267 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec1828e-1499-4e18-8b2d-5cfbe9a6d457-utilities\") pod \"redhat-operators-c4stk\" (UID: \"aec1828e-1499-4e18-8b2d-5cfbe9a6d457\") " pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.488328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vppw7\" (UniqueName: \"kubernetes.io/projected/aec1828e-1499-4e18-8b2d-5cfbe9a6d457-kube-api-access-vppw7\") pod \"redhat-operators-c4stk\" (UID: \"aec1828e-1499-4e18-8b2d-5cfbe9a6d457\") " pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.508476 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vppw7\" (UniqueName: \"kubernetes.io/projected/aec1828e-1499-4e18-8b2d-5cfbe9a6d457-kube-api-access-vppw7\") pod \"redhat-operators-c4stk\" (UID: \"aec1828e-1499-4e18-8b2d-5cfbe9a6d457\") " pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.666841 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.722358 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gvxkv"] Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.752020 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4x4r" event={"ID":"af33352e-df7a-4fb2-af44-d084a141e402","Type":"ContainerStarted","Data":"dbc238e24031a1c22746c2beabb413159ddb911bf617c7386152e288910cf307"} Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.758437 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkthb" event={"ID":"ceaf5522-aed5-418f-b483-4c0b945dc615","Type":"ContainerStarted","Data":"0633b3e586f3e265219f069598766011984576c91873833e798ab6bc36992114"} Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.759786 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" event={"ID":"67c85582-9e2c-43ed-9a9f-dc46ce546421","Type":"ContainerStarted","Data":"1b083c3e2ff00364d2446544cc4a3ba54188de7484beda5619cf84b0551f9e3f"} Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.759823 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" event={"ID":"67c85582-9e2c-43ed-9a9f-dc46ce546421","Type":"ContainerStarted","Data":"7c0a13885cef77766b831303a553c9a5eac5b8953b189b6193d68ad34afe9d5c"} Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.760792 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.765047 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.771855 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d4x4r" podStartSLOduration=2.285946139 podStartE2EDuration="3.771842345s" podCreationTimestamp="2026-02-15 20:23:00 +0000 UTC" firstStartedPulling="2026-02-15 20:23:01.692316007 +0000 UTC m=+389.558331630" lastFinishedPulling="2026-02-15 20:23:03.178212213 +0000 UTC m=+391.044227836" observedRunningTime="2026-02-15 20:23:03.770316444 +0000 UTC m=+391.636332087" watchObservedRunningTime="2026-02-15 20:23:03.771842345 +0000 UTC m=+391.637857968" Feb 15 20:23:03 crc kubenswrapper[4735]: I0215 20:23:03.795248 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7f44b59745-xszbh" podStartSLOduration=2.7952303130000002 podStartE2EDuration="2.795230313s" podCreationTimestamp="2026-02-15 20:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:23:03.791733977 +0000 UTC m=+391.657749600" watchObservedRunningTime="2026-02-15 20:23:03.795230313 +0000 UTC m=+391.661245936" Feb 15 20:23:04 crc kubenswrapper[4735]: I0215 20:23:04.053607 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4stk"] Feb 15 20:23:04 crc kubenswrapper[4735]: W0215 20:23:04.069391 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaec1828e_1499_4e18_8b2d_5cfbe9a6d457.slice/crio-485d54ab7e527344a102ee946a0c68436c5df42ffd41a1de07390aaf27b007ae WatchSource:0}: Error finding container 485d54ab7e527344a102ee946a0c68436c5df42ffd41a1de07390aaf27b007ae: Status 404 returned error can't find the container with id 485d54ab7e527344a102ee946a0c68436c5df42ffd41a1de07390aaf27b007ae Feb 15 20:23:04 crc kubenswrapper[4735]: I0215 20:23:04.770018 4735 generic.go:334] "Generic (PLEG): container finished" podID="aec1828e-1499-4e18-8b2d-5cfbe9a6d457" containerID="dd6cffbae2301cf8ce9a0d55880915cc5c0330a847c11f475e8c0f6b27a64ff0" exitCode=0 Feb 15 20:23:04 crc kubenswrapper[4735]: I0215 20:23:04.770076 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4stk" event={"ID":"aec1828e-1499-4e18-8b2d-5cfbe9a6d457","Type":"ContainerDied","Data":"dd6cffbae2301cf8ce9a0d55880915cc5c0330a847c11f475e8c0f6b27a64ff0"} Feb 15 20:23:04 crc kubenswrapper[4735]: I0215 20:23:04.771452 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4stk" event={"ID":"aec1828e-1499-4e18-8b2d-5cfbe9a6d457","Type":"ContainerStarted","Data":"485d54ab7e527344a102ee946a0c68436c5df42ffd41a1de07390aaf27b007ae"} Feb 15 20:23:04 crc kubenswrapper[4735]: I0215 20:23:04.772692 4735 generic.go:334] "Generic (PLEG): container finished" podID="f41f0acf-7b74-44f7-92a4-30fb79ccea7a" containerID="ac550e38b00468d5afca36decf9afdb0da0fcc1382f36db84ecd700753265522" exitCode=0 Feb 15 20:23:04 crc kubenswrapper[4735]: I0215 20:23:04.772786 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvxkv" event={"ID":"f41f0acf-7b74-44f7-92a4-30fb79ccea7a","Type":"ContainerDied","Data":"ac550e38b00468d5afca36decf9afdb0da0fcc1382f36db84ecd700753265522"} Feb 15 20:23:04 crc kubenswrapper[4735]: I0215 20:23:04.772810 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvxkv" event={"ID":"f41f0acf-7b74-44f7-92a4-30fb79ccea7a","Type":"ContainerStarted","Data":"41f13cbb3b81af5208025fb372a965b167fc0be0a999a0f3d6bf32f8eb7d2ded"} Feb 15 20:23:04 crc kubenswrapper[4735]: I0215 20:23:04.776042 4735 generic.go:334] "Generic (PLEG): container finished" podID="ceaf5522-aed5-418f-b483-4c0b945dc615" containerID="0633b3e586f3e265219f069598766011984576c91873833e798ab6bc36992114" exitCode=0 Feb 15 20:23:04 crc kubenswrapper[4735]: I0215 20:23:04.776825 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkthb" event={"ID":"ceaf5522-aed5-418f-b483-4c0b945dc615","Type":"ContainerDied","Data":"0633b3e586f3e265219f069598766011984576c91873833e798ab6bc36992114"} Feb 15 20:23:05 crc kubenswrapper[4735]: I0215 20:23:05.783090 4735 generic.go:334] "Generic (PLEG): container finished" podID="f41f0acf-7b74-44f7-92a4-30fb79ccea7a" containerID="ba6bace6efe1bcb95cb72049e6b6c08b15f7210fb74ffbd9b87edae6aafa598b" exitCode=0 Feb 15 20:23:05 crc kubenswrapper[4735]: I0215 20:23:05.783297 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvxkv" event={"ID":"f41f0acf-7b74-44f7-92a4-30fb79ccea7a","Type":"ContainerDied","Data":"ba6bace6efe1bcb95cb72049e6b6c08b15f7210fb74ffbd9b87edae6aafa598b"} Feb 15 20:23:05 crc kubenswrapper[4735]: I0215 20:23:05.786211 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkthb" event={"ID":"ceaf5522-aed5-418f-b483-4c0b945dc615","Type":"ContainerStarted","Data":"f1d50cd778cd701e50c4fee12c10dfbecdfae4318ee235c0d8b365e697732e18"} Feb 15 20:23:05 crc kubenswrapper[4735]: I0215 20:23:05.790560 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4stk" event={"ID":"aec1828e-1499-4e18-8b2d-5cfbe9a6d457","Type":"ContainerStarted","Data":"2b003dd02ed212197c7e3c627a5da29bfb7d339a142ae521a766f131f8f3cef5"} Feb 15 20:23:05 crc kubenswrapper[4735]: I0215 20:23:05.840788 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bkthb" podStartSLOduration=3.38056405 podStartE2EDuration="5.840767262s" podCreationTimestamp="2026-02-15 20:23:00 +0000 UTC" firstStartedPulling="2026-02-15 20:23:02.717973018 +0000 UTC m=+390.583988641" lastFinishedPulling="2026-02-15 20:23:05.17817623 +0000 UTC m=+393.044191853" observedRunningTime="2026-02-15 20:23:05.83737732 +0000 UTC m=+393.703392943" watchObservedRunningTime="2026-02-15 20:23:05.840767262 +0000 UTC m=+393.706782885" Feb 15 20:23:06 crc kubenswrapper[4735]: I0215 20:23:06.799450 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvxkv" event={"ID":"f41f0acf-7b74-44f7-92a4-30fb79ccea7a","Type":"ContainerStarted","Data":"57655164dbb70e17359b19fb60b212a9e65c5c26694ca7980d40b8653ba81b7a"} Feb 15 20:23:06 crc kubenswrapper[4735]: I0215 20:23:06.802074 4735 generic.go:334] "Generic (PLEG): container finished" podID="aec1828e-1499-4e18-8b2d-5cfbe9a6d457" containerID="2b003dd02ed212197c7e3c627a5da29bfb7d339a142ae521a766f131f8f3cef5" exitCode=0 Feb 15 20:23:06 crc kubenswrapper[4735]: I0215 20:23:06.802785 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4stk" event={"ID":"aec1828e-1499-4e18-8b2d-5cfbe9a6d457","Type":"ContainerDied","Data":"2b003dd02ed212197c7e3c627a5da29bfb7d339a142ae521a766f131f8f3cef5"} Feb 15 20:23:06 crc kubenswrapper[4735]: I0215 20:23:06.828190 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gvxkv" podStartSLOduration=2.4181886710000002 podStartE2EDuration="3.828175187s" podCreationTimestamp="2026-02-15 20:23:03 +0000 UTC" firstStartedPulling="2026-02-15 20:23:04.774567258 +0000 UTC m=+392.640582881" lastFinishedPulling="2026-02-15 20:23:06.184553774 +0000 UTC m=+394.050569397" observedRunningTime="2026-02-15 20:23:06.818918705 +0000 UTC m=+394.684934338" watchObservedRunningTime="2026-02-15 20:23:06.828175187 +0000 UTC m=+394.694190810" Feb 15 20:23:07 crc kubenswrapper[4735]: I0215 20:23:07.809062 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4stk" event={"ID":"aec1828e-1499-4e18-8b2d-5cfbe9a6d457","Type":"ContainerStarted","Data":"86df6bce172d5a167380a5cf6d477d93c23c6a8065de5c2eed8020684fbeb305"} Feb 15 20:23:07 crc kubenswrapper[4735]: I0215 20:23:07.831893 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c4stk" podStartSLOduration=2.441834688 podStartE2EDuration="4.831874678s" podCreationTimestamp="2026-02-15 20:23:03 +0000 UTC" firstStartedPulling="2026-02-15 20:23:04.776356887 +0000 UTC m=+392.642372510" lastFinishedPulling="2026-02-15 20:23:07.166396877 +0000 UTC m=+395.032412500" observedRunningTime="2026-02-15 20:23:07.830801588 +0000 UTC m=+395.696817211" watchObservedRunningTime="2026-02-15 20:23:07.831874678 +0000 UTC m=+395.697890571" Feb 15 20:23:08 crc kubenswrapper[4735]: E0215 20:23:08.018118 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/NetworkManager-dispatcher.service\": RecentStats: unable to find data in memory cache]" Feb 15 20:23:10 crc kubenswrapper[4735]: I0215 20:23:10.692416 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4zvg5" Feb 15 20:23:10 crc kubenswrapper[4735]: I0215 20:23:10.744200 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9h4lw"] Feb 15 20:23:11 crc kubenswrapper[4735]: I0215 20:23:11.072478 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:11 crc kubenswrapper[4735]: I0215 20:23:11.072519 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:11 crc kubenswrapper[4735]: I0215 20:23:11.122706 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:11 crc kubenswrapper[4735]: I0215 20:23:11.316899 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:11 crc kubenswrapper[4735]: I0215 20:23:11.317013 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:11 crc kubenswrapper[4735]: I0215 20:23:11.356667 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:11 crc kubenswrapper[4735]: I0215 20:23:11.893355 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d4x4r" Feb 15 20:23:11 crc kubenswrapper[4735]: I0215 20:23:11.899472 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bkthb" Feb 15 20:23:13 crc kubenswrapper[4735]: I0215 20:23:13.470345 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:13 crc kubenswrapper[4735]: I0215 20:23:13.470428 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:13 crc kubenswrapper[4735]: I0215 20:23:13.508673 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:13 crc kubenswrapper[4735]: I0215 20:23:13.667752 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:13 crc kubenswrapper[4735]: I0215 20:23:13.667814 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:13 crc kubenswrapper[4735]: I0215 20:23:13.701550 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:13 crc kubenswrapper[4735]: I0215 20:23:13.907166 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c4stk" Feb 15 20:23:13 crc kubenswrapper[4735]: I0215 20:23:13.920388 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gvxkv" Feb 15 20:23:19 crc kubenswrapper[4735]: I0215 20:23:19.679995 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:23:19 crc kubenswrapper[4735]: I0215 20:23:19.680559 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:23:35 crc kubenswrapper[4735]: I0215 20:23:35.790722 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" podUID="d4b1b306-2224-4908-b9f6-484d76a77ee1" containerName="registry" containerID="cri-o://3388046baa5114834e8c40f562b69fc131769c6832fa605d306b2911327e3e7a" gracePeriod=30 Feb 15 20:23:35 crc kubenswrapper[4735]: I0215 20:23:35.973671 4735 generic.go:334] "Generic (PLEG): container finished" podID="d4b1b306-2224-4908-b9f6-484d76a77ee1" containerID="3388046baa5114834e8c40f562b69fc131769c6832fa605d306b2911327e3e7a" exitCode=0 Feb 15 20:23:35 crc kubenswrapper[4735]: I0215 20:23:35.973734 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" event={"ID":"d4b1b306-2224-4908-b9f6-484d76a77ee1","Type":"ContainerDied","Data":"3388046baa5114834e8c40f562b69fc131769c6832fa605d306b2911327e3e7a"} Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.218210 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.280053 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-trusted-ca\") pod \"d4b1b306-2224-4908-b9f6-484d76a77ee1\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.280180 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-tls\") pod \"d4b1b306-2224-4908-b9f6-484d76a77ee1\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.280254 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-certificates\") pod \"d4b1b306-2224-4908-b9f6-484d76a77ee1\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.280303 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d4b1b306-2224-4908-b9f6-484d76a77ee1-installation-pull-secrets\") pod \"d4b1b306-2224-4908-b9f6-484d76a77ee1\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.280489 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d4b1b306-2224-4908-b9f6-484d76a77ee1\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.280529 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d4b1b306-2224-4908-b9f6-484d76a77ee1-ca-trust-extracted\") pod \"d4b1b306-2224-4908-b9f6-484d76a77ee1\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.280590 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtjzh\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-kube-api-access-rtjzh\") pod \"d4b1b306-2224-4908-b9f6-484d76a77ee1\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.280640 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-bound-sa-token\") pod \"d4b1b306-2224-4908-b9f6-484d76a77ee1\" (UID: \"d4b1b306-2224-4908-b9f6-484d76a77ee1\") " Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.284713 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d4b1b306-2224-4908-b9f6-484d76a77ee1" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.285393 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d4b1b306-2224-4908-b9f6-484d76a77ee1" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.297339 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d4b1b306-2224-4908-b9f6-484d76a77ee1" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.299404 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d4b1b306-2224-4908-b9f6-484d76a77ee1" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.300256 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4b1b306-2224-4908-b9f6-484d76a77ee1-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d4b1b306-2224-4908-b9f6-484d76a77ee1" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.307455 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d4b1b306-2224-4908-b9f6-484d76a77ee1" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.307682 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-kube-api-access-rtjzh" (OuterVolumeSpecName: "kube-api-access-rtjzh") pod "d4b1b306-2224-4908-b9f6-484d76a77ee1" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1"). InnerVolumeSpecName "kube-api-access-rtjzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.311133 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4b1b306-2224-4908-b9f6-484d76a77ee1-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d4b1b306-2224-4908-b9f6-484d76a77ee1" (UID: "d4b1b306-2224-4908-b9f6-484d76a77ee1"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.382981 4735 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d4b1b306-2224-4908-b9f6-484d76a77ee1-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.383040 4735 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d4b1b306-2224-4908-b9f6-484d76a77ee1-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.383060 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtjzh\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-kube-api-access-rtjzh\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.383078 4735 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.383099 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.383116 4735 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.383134 4735 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d4b1b306-2224-4908-b9f6-484d76a77ee1-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.980885 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" event={"ID":"d4b1b306-2224-4908-b9f6-484d76a77ee1","Type":"ContainerDied","Data":"0b832c4933cf53e1ea10ce401b97a61c2a784e7be0b9621036f68dcaa5fafd7d"} Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.980915 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9h4lw" Feb 15 20:23:36 crc kubenswrapper[4735]: I0215 20:23:36.980981 4735 scope.go:117] "RemoveContainer" containerID="3388046baa5114834e8c40f562b69fc131769c6832fa605d306b2911327e3e7a" Feb 15 20:23:37 crc kubenswrapper[4735]: I0215 20:23:37.009164 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9h4lw"] Feb 15 20:23:37 crc kubenswrapper[4735]: I0215 20:23:37.016210 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9h4lw"] Feb 15 20:23:38 crc kubenswrapper[4735]: I0215 20:23:38.892838 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4b1b306-2224-4908-b9f6-484d76a77ee1" path="/var/lib/kubelet/pods/d4b1b306-2224-4908-b9f6-484d76a77ee1/volumes" Feb 15 20:23:49 crc kubenswrapper[4735]: I0215 20:23:49.680051 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:23:49 crc kubenswrapper[4735]: I0215 20:23:49.680801 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:23:49 crc kubenswrapper[4735]: I0215 20:23:49.680880 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:23:49 crc kubenswrapper[4735]: I0215 20:23:49.681842 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"922ecbdcfc9b4bcb4d36975920d88c0963942c6e49ace39850b6e39e406cc403"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:23:49 crc kubenswrapper[4735]: I0215 20:23:49.682071 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://922ecbdcfc9b4bcb4d36975920d88c0963942c6e49ace39850b6e39e406cc403" gracePeriod=600 Feb 15 20:23:50 crc kubenswrapper[4735]: I0215 20:23:50.054408 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="922ecbdcfc9b4bcb4d36975920d88c0963942c6e49ace39850b6e39e406cc403" exitCode=0 Feb 15 20:23:50 crc kubenswrapper[4735]: I0215 20:23:50.054671 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"922ecbdcfc9b4bcb4d36975920d88c0963942c6e49ace39850b6e39e406cc403"} Feb 15 20:23:50 crc kubenswrapper[4735]: I0215 20:23:50.054747 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"925e7518253cb6621d6125aa829f6d9222f7f5dddb79f1eba385c5daf1d88f7d"} Feb 15 20:23:50 crc kubenswrapper[4735]: I0215 20:23:50.054766 4735 scope.go:117] "RemoveContainer" containerID="bdca5e79ba04d90058327b5d21ddd8277a6e08a36783bbe984d91699f72bbb28" Feb 15 20:25:49 crc kubenswrapper[4735]: I0215 20:25:49.680602 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:25:49 crc kubenswrapper[4735]: I0215 20:25:49.681434 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:26:19 crc kubenswrapper[4735]: I0215 20:26:19.679894 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:26:19 crc kubenswrapper[4735]: I0215 20:26:19.680514 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:26:49 crc kubenswrapper[4735]: I0215 20:26:49.680569 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:26:49 crc kubenswrapper[4735]: I0215 20:26:49.681272 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:26:49 crc kubenswrapper[4735]: I0215 20:26:49.681358 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:26:49 crc kubenswrapper[4735]: I0215 20:26:49.682384 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"925e7518253cb6621d6125aa829f6d9222f7f5dddb79f1eba385c5daf1d88f7d"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:26:49 crc kubenswrapper[4735]: I0215 20:26:49.682471 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://925e7518253cb6621d6125aa829f6d9222f7f5dddb79f1eba385c5daf1d88f7d" gracePeriod=600 Feb 15 20:26:50 crc kubenswrapper[4735]: I0215 20:26:50.469667 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="925e7518253cb6621d6125aa829f6d9222f7f5dddb79f1eba385c5daf1d88f7d" exitCode=0 Feb 15 20:26:50 crc kubenswrapper[4735]: I0215 20:26:50.469745 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"925e7518253cb6621d6125aa829f6d9222f7f5dddb79f1eba385c5daf1d88f7d"} Feb 15 20:26:50 crc kubenswrapper[4735]: I0215 20:26:50.470209 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"8137201feee74b50c12e4ac603bd7a608ffb1bf6f8ae2e8df072ae67e560dfe2"} Feb 15 20:26:50 crc kubenswrapper[4735]: I0215 20:26:50.470227 4735 scope.go:117] "RemoveContainer" containerID="922ecbdcfc9b4bcb4d36975920d88c0963942c6e49ace39850b6e39e406cc403" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.140474 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-88v64"] Feb 15 20:27:46 crc kubenswrapper[4735]: E0215 20:27:46.141341 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4b1b306-2224-4908-b9f6-484d76a77ee1" containerName="registry" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.141362 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4b1b306-2224-4908-b9f6-484d76a77ee1" containerName="registry" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.141529 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4b1b306-2224-4908-b9f6-484d76a77ee1" containerName="registry" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.142199 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-88v64" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.144283 4735 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-7lvhj" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.145597 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.146601 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-88v64"] Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.148628 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.186077 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-2jh8n"] Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.187722 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-2jh8n" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.191632 4735 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jv8k9" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.202234 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-jvntw"] Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.202970 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.205408 4735 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-9b92q" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.206434 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-2jh8n"] Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.212647 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-jvntw"] Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.272391 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kphz5\" (UniqueName: \"kubernetes.io/projected/468178fa-6b3c-41b4-9bf2-45178ebfb3c2-kube-api-access-kphz5\") pod \"cert-manager-webhook-687f57d79b-jvntw\" (UID: \"468178fa-6b3c-41b4-9bf2-45178ebfb3c2\") " pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.272459 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghkl5\" (UniqueName: \"kubernetes.io/projected/e04e2b12-a2ee-4ba4-acb4-0276c777a066-kube-api-access-ghkl5\") pod \"cert-manager-cainjector-cf98fcc89-88v64\" (UID: \"e04e2b12-a2ee-4ba4-acb4-0276c777a066\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-88v64" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.272487 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h95j4\" (UniqueName: \"kubernetes.io/projected/56f25e9b-225b-433d-af08-bdd6bb37e6e4-kube-api-access-h95j4\") pod \"cert-manager-858654f9db-2jh8n\" (UID: \"56f25e9b-225b-433d-af08-bdd6bb37e6e4\") " pod="cert-manager/cert-manager-858654f9db-2jh8n" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.374041 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kphz5\" (UniqueName: \"kubernetes.io/projected/468178fa-6b3c-41b4-9bf2-45178ebfb3c2-kube-api-access-kphz5\") pod \"cert-manager-webhook-687f57d79b-jvntw\" (UID: \"468178fa-6b3c-41b4-9bf2-45178ebfb3c2\") " pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.374128 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghkl5\" (UniqueName: \"kubernetes.io/projected/e04e2b12-a2ee-4ba4-acb4-0276c777a066-kube-api-access-ghkl5\") pod \"cert-manager-cainjector-cf98fcc89-88v64\" (UID: \"e04e2b12-a2ee-4ba4-acb4-0276c777a066\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-88v64" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.374164 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h95j4\" (UniqueName: \"kubernetes.io/projected/56f25e9b-225b-433d-af08-bdd6bb37e6e4-kube-api-access-h95j4\") pod \"cert-manager-858654f9db-2jh8n\" (UID: \"56f25e9b-225b-433d-af08-bdd6bb37e6e4\") " pod="cert-manager/cert-manager-858654f9db-2jh8n" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.403668 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h95j4\" (UniqueName: \"kubernetes.io/projected/56f25e9b-225b-433d-af08-bdd6bb37e6e4-kube-api-access-h95j4\") pod \"cert-manager-858654f9db-2jh8n\" (UID: \"56f25e9b-225b-433d-af08-bdd6bb37e6e4\") " pod="cert-manager/cert-manager-858654f9db-2jh8n" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.403716 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghkl5\" (UniqueName: \"kubernetes.io/projected/e04e2b12-a2ee-4ba4-acb4-0276c777a066-kube-api-access-ghkl5\") pod \"cert-manager-cainjector-cf98fcc89-88v64\" (UID: \"e04e2b12-a2ee-4ba4-acb4-0276c777a066\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-88v64" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.403810 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kphz5\" (UniqueName: \"kubernetes.io/projected/468178fa-6b3c-41b4-9bf2-45178ebfb3c2-kube-api-access-kphz5\") pod \"cert-manager-webhook-687f57d79b-jvntw\" (UID: \"468178fa-6b3c-41b4-9bf2-45178ebfb3c2\") " pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.463698 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-88v64" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.521502 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-2jh8n" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.529494 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.924877 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-88v64"] Feb 15 20:27:46 crc kubenswrapper[4735]: W0215 20:27:46.929209 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode04e2b12_a2ee_4ba4_acb4_0276c777a066.slice/crio-f1a864326751867aa650d194a6ce7c10753c9fcb81102ce07152b87bee42d258 WatchSource:0}: Error finding container f1a864326751867aa650d194a6ce7c10753c9fcb81102ce07152b87bee42d258: Status 404 returned error can't find the container with id f1a864326751867aa650d194a6ce7c10753c9fcb81102ce07152b87bee42d258 Feb 15 20:27:46 crc kubenswrapper[4735]: I0215 20:27:46.931358 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 20:27:47 crc kubenswrapper[4735]: I0215 20:27:47.000026 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-2jh8n"] Feb 15 20:27:47 crc kubenswrapper[4735]: W0215 20:27:47.006598 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56f25e9b_225b_433d_af08_bdd6bb37e6e4.slice/crio-9a0f0adac99781f3f16d7c2119bc53b18a0fe431855465621921bb7e36986c46 WatchSource:0}: Error finding container 9a0f0adac99781f3f16d7c2119bc53b18a0fe431855465621921bb7e36986c46: Status 404 returned error can't find the container with id 9a0f0adac99781f3f16d7c2119bc53b18a0fe431855465621921bb7e36986c46 Feb 15 20:27:47 crc kubenswrapper[4735]: I0215 20:27:47.045772 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-jvntw"] Feb 15 20:27:47 crc kubenswrapper[4735]: W0215 20:27:47.049250 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod468178fa_6b3c_41b4_9bf2_45178ebfb3c2.slice/crio-68f5a3d0161b9734018516ce60afd0239ea4dc712e8bdfec508f799b31ba9fbb WatchSource:0}: Error finding container 68f5a3d0161b9734018516ce60afd0239ea4dc712e8bdfec508f799b31ba9fbb: Status 404 returned error can't find the container with id 68f5a3d0161b9734018516ce60afd0239ea4dc712e8bdfec508f799b31ba9fbb Feb 15 20:27:47 crc kubenswrapper[4735]: I0215 20:27:47.824250 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" event={"ID":"468178fa-6b3c-41b4-9bf2-45178ebfb3c2","Type":"ContainerStarted","Data":"68f5a3d0161b9734018516ce60afd0239ea4dc712e8bdfec508f799b31ba9fbb"} Feb 15 20:27:47 crc kubenswrapper[4735]: I0215 20:27:47.825813 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-2jh8n" event={"ID":"56f25e9b-225b-433d-af08-bdd6bb37e6e4","Type":"ContainerStarted","Data":"9a0f0adac99781f3f16d7c2119bc53b18a0fe431855465621921bb7e36986c46"} Feb 15 20:27:47 crc kubenswrapper[4735]: I0215 20:27:47.827180 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-88v64" event={"ID":"e04e2b12-a2ee-4ba4-acb4-0276c777a066","Type":"ContainerStarted","Data":"f1a864326751867aa650d194a6ce7c10753c9fcb81102ce07152b87bee42d258"} Feb 15 20:27:49 crc kubenswrapper[4735]: I0215 20:27:49.840048 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" event={"ID":"468178fa-6b3c-41b4-9bf2-45178ebfb3c2","Type":"ContainerStarted","Data":"0bd8e1316b17d05f3b844a34c5af3f61bc3b1f528cac44da1299ad0168ebcac6"} Feb 15 20:27:49 crc kubenswrapper[4735]: I0215 20:27:49.840606 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" Feb 15 20:27:49 crc kubenswrapper[4735]: I0215 20:27:49.858552 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" podStartSLOduration=1.23812143 podStartE2EDuration="3.85852203s" podCreationTimestamp="2026-02-15 20:27:46 +0000 UTC" firstStartedPulling="2026-02-15 20:27:47.05303792 +0000 UTC m=+674.919053533" lastFinishedPulling="2026-02-15 20:27:49.6734385 +0000 UTC m=+677.539454133" observedRunningTime="2026-02-15 20:27:49.853094523 +0000 UTC m=+677.719110166" watchObservedRunningTime="2026-02-15 20:27:49.85852203 +0000 UTC m=+677.724537663" Feb 15 20:27:50 crc kubenswrapper[4735]: I0215 20:27:50.847007 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-88v64" event={"ID":"e04e2b12-a2ee-4ba4-acb4-0276c777a066","Type":"ContainerStarted","Data":"2fb073833ee6dca841dd5c1b35f8e2c13882c7ca7e4026eabcdac05e98c3c976"} Feb 15 20:27:50 crc kubenswrapper[4735]: I0215 20:27:50.870177 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-88v64" podStartSLOduration=2.075268322 podStartE2EDuration="4.870146204s" podCreationTimestamp="2026-02-15 20:27:46 +0000 UTC" firstStartedPulling="2026-02-15 20:27:46.931090983 +0000 UTC m=+674.797106616" lastFinishedPulling="2026-02-15 20:27:49.725968875 +0000 UTC m=+677.591984498" observedRunningTime="2026-02-15 20:27:50.861789099 +0000 UTC m=+678.727804732" watchObservedRunningTime="2026-02-15 20:27:50.870146204 +0000 UTC m=+678.736161867" Feb 15 20:27:51 crc kubenswrapper[4735]: I0215 20:27:51.857352 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-2jh8n" event={"ID":"56f25e9b-225b-433d-af08-bdd6bb37e6e4","Type":"ContainerStarted","Data":"dc4c76ba3803d9937cbca6073d382ff6fe7b17c5fca3a4d20f9c5a016d584cc5"} Feb 15 20:27:51 crc kubenswrapper[4735]: I0215 20:27:51.880197 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-2jh8n" podStartSLOduration=2.112877094 podStartE2EDuration="5.880175815s" podCreationTimestamp="2026-02-15 20:27:46 +0000 UTC" firstStartedPulling="2026-02-15 20:27:47.008884689 +0000 UTC m=+674.874900322" lastFinishedPulling="2026-02-15 20:27:50.77618342 +0000 UTC m=+678.642199043" observedRunningTime="2026-02-15 20:27:51.874211754 +0000 UTC m=+679.740227377" watchObservedRunningTime="2026-02-15 20:27:51.880175815 +0000 UTC m=+679.746191448" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.451547 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x9xmf"] Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.452305 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovn-controller" containerID="cri-o://d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62" gracePeriod=30 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.452383 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="nbdb" containerID="cri-o://88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920" gracePeriod=30 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.452446 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55" gracePeriod=30 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.452491 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kube-rbac-proxy-node" containerID="cri-o://15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e" gracePeriod=30 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.452525 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovn-acl-logging" containerID="cri-o://50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc" gracePeriod=30 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.452432 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="northd" containerID="cri-o://3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0" gracePeriod=30 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.452705 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="sbdb" containerID="cri-o://498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec" gracePeriod=30 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.499895 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" containerID="cri-o://94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc" gracePeriod=30 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.532457 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.898057 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/2.log" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.899279 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/1.log" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.899482 4735 generic.go:334] "Generic (PLEG): container finished" podID="a20277cf-39a2-453f-ac6d-fb28346b3358" containerID="987c4a7181b247e73c933f1f073f70911e9100019abf6fd2a37c154a601e5c62" exitCode=2 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.899565 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wfjmx" event={"ID":"a20277cf-39a2-453f-ac6d-fb28346b3358","Type":"ContainerDied","Data":"987c4a7181b247e73c933f1f073f70911e9100019abf6fd2a37c154a601e5c62"} Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.899830 4735 scope.go:117] "RemoveContainer" containerID="17e0edc4d80b21ff16e632dd631d7b3fdbed7e4179ed0c98c17fd921b94c3889" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.900755 4735 scope.go:117] "RemoveContainer" containerID="987c4a7181b247e73c933f1f073f70911e9100019abf6fd2a37c154a601e5c62" Feb 15 20:27:56 crc kubenswrapper[4735]: E0215 20:27:56.900918 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-wfjmx_openshift-multus(a20277cf-39a2-453f-ac6d-fb28346b3358)\"" pod="openshift-multus/multus-wfjmx" podUID="a20277cf-39a2-453f-ac6d-fb28346b3358" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.902548 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovnkube-controller/3.log" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.909878 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovn-acl-logging/0.log" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.910434 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovn-controller/0.log" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.910987 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc" exitCode=0 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.911399 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc"} Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.911490 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec" exitCode=0 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912229 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920" exitCode=0 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912198 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec"} Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912301 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920"} Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912326 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55"} Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912243 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55" exitCode=0 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912379 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e" exitCode=0 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912410 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc" exitCode=143 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912427 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62" exitCode=143 Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912452 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e"} Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912528 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc"} Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.912582 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62"} Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.927623 4735 scope.go:117] "RemoveContainer" containerID="dbaae250762e20f85f66583c2b804ead54e1cc4d174a5ac0606a91484f7d8d3f" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.949615 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovn-acl-logging/0.log" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.950382 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovn-controller/0.log" Feb 15 20:27:56 crc kubenswrapper[4735]: I0215 20:27:56.950919 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017317 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nczmn"] Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017559 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017572 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017582 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovn-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017589 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovn-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017602 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="nbdb" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017610 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="nbdb" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017623 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="sbdb" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017633 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="sbdb" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017643 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovn-acl-logging" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017649 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovn-acl-logging" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017659 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017665 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017674 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kube-rbac-proxy-node" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017680 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kube-rbac-proxy-node" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017686 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="northd" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017693 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="northd" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017700 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017706 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017713 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kubecfg-setup" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017719 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kubecfg-setup" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017729 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017736 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.017744 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kube-rbac-proxy-ovn-metrics" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017753 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kube-rbac-proxy-ovn-metrics" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017846 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovn-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017853 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="nbdb" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017863 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="northd" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017872 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kube-rbac-proxy-ovn-metrics" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017877 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017884 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovn-acl-logging" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017891 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="kube-rbac-proxy-node" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017897 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017907 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017916 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017923 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.017930 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="sbdb" Feb 15 20:27:57 crc kubenswrapper[4735]: E0215 20:27:57.018036 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.018045 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerName="ovnkube-controller" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.019563 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038279 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-script-lib\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038331 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-openvswitch\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038364 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-bin\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038385 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-config\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038406 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-node-log\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038429 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-var-lib-openvswitch\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038448 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-etc-openvswitch\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038509 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-netd\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038538 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-systemd-units\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038565 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-env-overrides\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038589 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-ovn-kubernetes\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038616 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-ovn\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038645 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk9bx\" (UniqueName: \"kubernetes.io/projected/755c2f64-a91b-443f-96ca-88c8cab0d656-kube-api-access-nk9bx\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038670 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-log-socket\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038711 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-kubelet\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038746 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-netns\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038774 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-slash\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038804 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/755c2f64-a91b-443f-96ca-88c8cab0d656-ovn-node-metrics-cert\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038835 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-systemd\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038858 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-var-lib-cni-networks-ovn-kubernetes\") pod \"755c2f64-a91b-443f-96ca-88c8cab0d656\" (UID: \"755c2f64-a91b-443f-96ca-88c8cab0d656\") " Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.038887 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039203 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039245 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039287 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039332 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039355 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039375 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039397 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039416 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039211 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-node-log" (OuterVolumeSpecName: "node-log") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.039710 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.040679 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-log-socket" (OuterVolumeSpecName: "log-socket") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.040718 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.040745 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.041035 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.041128 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.041126 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.041217 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-slash" (OuterVolumeSpecName: "host-slash") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.051415 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/755c2f64-a91b-443f-96ca-88c8cab0d656-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.054393 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/755c2f64-a91b-443f-96ca-88c8cab0d656-kube-api-access-nk9bx" (OuterVolumeSpecName: "kube-api-access-nk9bx") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "kube-api-access-nk9bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.066027 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "755c2f64-a91b-443f-96ca-88c8cab0d656" (UID: "755c2f64-a91b-443f-96ca-88c8cab0d656"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.140529 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-run-openvswitch\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.140574 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66bee24d-c697-413f-98ed-dd9e87163766-ovn-node-metrics-cert\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.140769 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.140876 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-cni-netd\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141057 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-systemd-units\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141172 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-slash\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141215 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-run-systemd\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141259 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-log-socket\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141296 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbqj2\" (UniqueName: \"kubernetes.io/projected/66bee24d-c697-413f-98ed-dd9e87163766-kube-api-access-cbqj2\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141400 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-node-log\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141441 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-run-netns\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141483 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-etc-openvswitch\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141548 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-kubelet\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141602 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-cni-bin\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141737 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66bee24d-c697-413f-98ed-dd9e87163766-ovnkube-config\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141788 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-run-ovn-kubernetes\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.141879 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/66bee24d-c697-413f-98ed-dd9e87163766-ovnkube-script-lib\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142143 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-run-ovn\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142286 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-var-lib-openvswitch\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142334 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66bee24d-c697-413f-98ed-dd9e87163766-env-overrides\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142498 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/755c2f64-a91b-443f-96ca-88c8cab0d656-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142532 4735 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142555 4735 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142578 4735 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142597 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142645 4735 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142665 4735 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-node-log\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142683 4735 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142701 4735 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142722 4735 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142741 4735 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142759 4735 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/755c2f64-a91b-443f-96ca-88c8cab0d656-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142778 4735 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142796 4735 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142813 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk9bx\" (UniqueName: \"kubernetes.io/projected/755c2f64-a91b-443f-96ca-88c8cab0d656-kube-api-access-nk9bx\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142834 4735 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-log-socket\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142861 4735 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142887 4735 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.142915 4735 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/755c2f64-a91b-443f-96ca-88c8cab0d656-host-slash\") on node \"crc\" DevicePath \"\"" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.244177 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-var-lib-openvswitch\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.244353 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-var-lib-openvswitch\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.244494 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-run-ovn\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.244686 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-run-ovn\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.244683 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66bee24d-c697-413f-98ed-dd9e87163766-env-overrides\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.244884 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-run-openvswitch\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245059 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66bee24d-c697-413f-98ed-dd9e87163766-ovn-node-metrics-cert\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245184 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245277 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-cni-netd\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245360 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-cni-netd\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245057 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-run-openvswitch\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245384 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-systemd-units\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245319 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245497 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-slash\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245481 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-slash\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245679 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-systemd-units\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245700 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-run-systemd\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245857 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-log-socket\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245712 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66bee24d-c697-413f-98ed-dd9e87163766-env-overrides\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.245785 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-run-systemd\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246034 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-log-socket\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246127 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbqj2\" (UniqueName: \"kubernetes.io/projected/66bee24d-c697-413f-98ed-dd9e87163766-kube-api-access-cbqj2\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246245 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-etc-openvswitch\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246333 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-node-log\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246430 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-run-netns\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246562 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-kubelet\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246665 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-kubelet\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246329 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-etc-openvswitch\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246375 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-node-log\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246668 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-cni-bin\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246800 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66bee24d-c697-413f-98ed-dd9e87163766-ovnkube-config\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246829 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-run-ovn-kubernetes\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246511 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-run-netns\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246866 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/66bee24d-c697-413f-98ed-dd9e87163766-ovnkube-script-lib\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.246922 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-run-ovn-kubernetes\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.247115 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66bee24d-c697-413f-98ed-dd9e87163766-host-cni-bin\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.247646 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/66bee24d-c697-413f-98ed-dd9e87163766-ovnkube-script-lib\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.247829 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66bee24d-c697-413f-98ed-dd9e87163766-ovnkube-config\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.251916 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66bee24d-c697-413f-98ed-dd9e87163766-ovn-node-metrics-cert\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.266344 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbqj2\" (UniqueName: \"kubernetes.io/projected/66bee24d-c697-413f-98ed-dd9e87163766-kube-api-access-cbqj2\") pod \"ovnkube-node-nczmn\" (UID: \"66bee24d-c697-413f-98ed-dd9e87163766\") " pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.336901 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.934667 4735 generic.go:334] "Generic (PLEG): container finished" podID="66bee24d-c697-413f-98ed-dd9e87163766" containerID="acca8e900c606826731d2cefd7fec80c23ce8b7c99b3d8d8229fca6fbbb813dd" exitCode=0 Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.936466 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerDied","Data":"acca8e900c606826731d2cefd7fec80c23ce8b7c99b3d8d8229fca6fbbb813dd"} Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.936657 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerStarted","Data":"7c1ae065ad1c71dffb10dd59145dec25d468a3219cfc01b0347ec83054ae2011"} Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.939544 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/2.log" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.951001 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovn-acl-logging/0.log" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.951556 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x9xmf_755c2f64-a91b-443f-96ca-88c8cab0d656/ovn-controller/0.log" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.952125 4735 generic.go:334] "Generic (PLEG): container finished" podID="755c2f64-a91b-443f-96ca-88c8cab0d656" containerID="3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0" exitCode=0 Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.952184 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0"} Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.952214 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" event={"ID":"755c2f64-a91b-443f-96ca-88c8cab0d656","Type":"ContainerDied","Data":"2d31391d2ab043f7a2e016ed1ae052f7e5beead951464715e04f361c04e09a1d"} Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.952236 4735 scope.go:117] "RemoveContainer" containerID="94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.952415 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x9xmf" Feb 15 20:27:57 crc kubenswrapper[4735]: I0215 20:27:57.991422 4735 scope.go:117] "RemoveContainer" containerID="498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.035778 4735 scope.go:117] "RemoveContainer" containerID="88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.040185 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x9xmf"] Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.046308 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x9xmf"] Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.057983 4735 scope.go:117] "RemoveContainer" containerID="3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.075234 4735 scope.go:117] "RemoveContainer" containerID="ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.101060 4735 scope.go:117] "RemoveContainer" containerID="15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.136734 4735 scope.go:117] "RemoveContainer" containerID="50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.173728 4735 scope.go:117] "RemoveContainer" containerID="d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.199056 4735 scope.go:117] "RemoveContainer" containerID="a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.223805 4735 scope.go:117] "RemoveContainer" containerID="94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc" Feb 15 20:27:58 crc kubenswrapper[4735]: E0215 20:27:58.224254 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc\": container with ID starting with 94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc not found: ID does not exist" containerID="94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.224291 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc"} err="failed to get container status \"94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc\": rpc error: code = NotFound desc = could not find container \"94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc\": container with ID starting with 94f4822b34e6fe359bb78f74d7271e3c7701d36f8e58c4aca71b2d6148ba81dc not found: ID does not exist" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.224315 4735 scope.go:117] "RemoveContainer" containerID="498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec" Feb 15 20:27:58 crc kubenswrapper[4735]: E0215 20:27:58.224574 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\": container with ID starting with 498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec not found: ID does not exist" containerID="498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.224598 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec"} err="failed to get container status \"498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\": rpc error: code = NotFound desc = could not find container \"498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec\": container with ID starting with 498c16a666a9193513afb9562974d376b823ecfcde88c25b72f0d422dd901aec not found: ID does not exist" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.224618 4735 scope.go:117] "RemoveContainer" containerID="88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920" Feb 15 20:27:58 crc kubenswrapper[4735]: E0215 20:27:58.224881 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\": container with ID starting with 88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920 not found: ID does not exist" containerID="88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.224904 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920"} err="failed to get container status \"88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\": rpc error: code = NotFound desc = could not find container \"88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920\": container with ID starting with 88447b799034341a89fa4f8f4ea52cace26cc5b873c41d4128cf3aab6bb86920 not found: ID does not exist" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.224919 4735 scope.go:117] "RemoveContainer" containerID="3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0" Feb 15 20:27:58 crc kubenswrapper[4735]: E0215 20:27:58.225212 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\": container with ID starting with 3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0 not found: ID does not exist" containerID="3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.225232 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0"} err="failed to get container status \"3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\": rpc error: code = NotFound desc = could not find container \"3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0\": container with ID starting with 3662a2baef570d6359c2f2d3408e03075bed5076d37aae58a7c73d07b7885dd0 not found: ID does not exist" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.225248 4735 scope.go:117] "RemoveContainer" containerID="ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55" Feb 15 20:27:58 crc kubenswrapper[4735]: E0215 20:27:58.225490 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\": container with ID starting with ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55 not found: ID does not exist" containerID="ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.225512 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55"} err="failed to get container status \"ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\": rpc error: code = NotFound desc = could not find container \"ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55\": container with ID starting with ef713cfeea973232c2f55d9a5721c7f42231a4988b83efe055bc8e97a317ab55 not found: ID does not exist" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.225528 4735 scope.go:117] "RemoveContainer" containerID="15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e" Feb 15 20:27:58 crc kubenswrapper[4735]: E0215 20:27:58.225788 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\": container with ID starting with 15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e not found: ID does not exist" containerID="15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.225810 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e"} err="failed to get container status \"15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\": rpc error: code = NotFound desc = could not find container \"15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e\": container with ID starting with 15582e62800e9fdba16b570b5874963b92520941c5083928b6f98e6c058ab72e not found: ID does not exist" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.225827 4735 scope.go:117] "RemoveContainer" containerID="50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc" Feb 15 20:27:58 crc kubenswrapper[4735]: E0215 20:27:58.226342 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\": container with ID starting with 50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc not found: ID does not exist" containerID="50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.226362 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc"} err="failed to get container status \"50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\": rpc error: code = NotFound desc = could not find container \"50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc\": container with ID starting with 50291952a40676e6d0b7dae01deb8bb76cdbb1e34e1285d40933fa21cb7cd1cc not found: ID does not exist" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.226380 4735 scope.go:117] "RemoveContainer" containerID="d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62" Feb 15 20:27:58 crc kubenswrapper[4735]: E0215 20:27:58.226640 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\": container with ID starting with d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62 not found: ID does not exist" containerID="d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.226660 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62"} err="failed to get container status \"d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\": rpc error: code = NotFound desc = could not find container \"d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62\": container with ID starting with d08ddfdb73d3e0a6496ce8dab82b730aac9a213a717ca53684d635686abd4b62 not found: ID does not exist" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.226678 4735 scope.go:117] "RemoveContainer" containerID="a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48" Feb 15 20:27:58 crc kubenswrapper[4735]: E0215 20:27:58.226998 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\": container with ID starting with a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48 not found: ID does not exist" containerID="a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.227043 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48"} err="failed to get container status \"a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\": rpc error: code = NotFound desc = could not find container \"a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48\": container with ID starting with a4223d827ac9e499f42f097272153367129272fbdecfdb9def3536b8abd4ff48 not found: ID does not exist" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.896097 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="755c2f64-a91b-443f-96ca-88c8cab0d656" path="/var/lib/kubelet/pods/755c2f64-a91b-443f-96ca-88c8cab0d656/volumes" Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.969414 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerStarted","Data":"fc95ef5cbb3c9a99add6e9213920898a3c4dc483d23ec19572d24c801d1a67b0"} Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.969481 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerStarted","Data":"e793387c9b572ad42636622c2cc1dc6f06c58fd49c95fdfc87922956883f35b0"} Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.969508 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerStarted","Data":"33a168105b999105c8c89d2e6b7479ee2a0148c717ea13a1e8854ff10f3d8df7"} Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.969531 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerStarted","Data":"edb3b0d4f3f3fa367822f0d3480097b40fce1a0b1d2d3e079b7b2427977b493b"} Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.969553 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerStarted","Data":"0c47f2c26da246f2f831463bea4ec4b28546577d2063ea54e8b6f8820fa4240f"} Feb 15 20:27:58 crc kubenswrapper[4735]: I0215 20:27:58.969570 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerStarted","Data":"5335b5ec9a473651995d87c53d452ebd082456ee2998decd7d48a1141ff3e3cb"} Feb 15 20:28:02 crc kubenswrapper[4735]: I0215 20:28:02.003869 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerStarted","Data":"398083380be6673f32b0c0c748ce587942fc5a29ce3df33b235eaedfadddacb2"} Feb 15 20:28:04 crc kubenswrapper[4735]: I0215 20:28:04.016270 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" event={"ID":"66bee24d-c697-413f-98ed-dd9e87163766","Type":"ContainerStarted","Data":"d5e58f51ed9440cce93a92c9ca494a3489dbb6261951235d48273bef1d463d37"} Feb 15 20:28:04 crc kubenswrapper[4735]: I0215 20:28:04.016909 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:28:04 crc kubenswrapper[4735]: I0215 20:28:04.016928 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:28:04 crc kubenswrapper[4735]: I0215 20:28:04.042373 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:28:04 crc kubenswrapper[4735]: I0215 20:28:04.057173 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" podStartSLOduration=7.057154663 podStartE2EDuration="7.057154663s" podCreationTimestamp="2026-02-15 20:27:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:28:04.051337426 +0000 UTC m=+691.917353049" watchObservedRunningTime="2026-02-15 20:28:04.057154663 +0000 UTC m=+691.923170286" Feb 15 20:28:05 crc kubenswrapper[4735]: I0215 20:28:05.022625 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:28:05 crc kubenswrapper[4735]: I0215 20:28:05.064293 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:28:11 crc kubenswrapper[4735]: I0215 20:28:11.887174 4735 scope.go:117] "RemoveContainer" containerID="987c4a7181b247e73c933f1f073f70911e9100019abf6fd2a37c154a601e5c62" Feb 15 20:28:11 crc kubenswrapper[4735]: E0215 20:28:11.888199 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-wfjmx_openshift-multus(a20277cf-39a2-453f-ac6d-fb28346b3358)\"" pod="openshift-multus/multus-wfjmx" podUID="a20277cf-39a2-453f-ac6d-fb28346b3358" Feb 15 20:28:22 crc kubenswrapper[4735]: I0215 20:28:22.893325 4735 scope.go:117] "RemoveContainer" containerID="987c4a7181b247e73c933f1f073f70911e9100019abf6fd2a37c154a601e5c62" Feb 15 20:28:24 crc kubenswrapper[4735]: I0215 20:28:24.146502 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wfjmx_a20277cf-39a2-453f-ac6d-fb28346b3358/kube-multus/2.log" Feb 15 20:28:24 crc kubenswrapper[4735]: I0215 20:28:24.148755 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wfjmx" event={"ID":"a20277cf-39a2-453f-ac6d-fb28346b3358","Type":"ContainerStarted","Data":"1883ba5196e0b512c7da5146c21d26a068564f2eea07a76c5dfff9b14fe56a46"} Feb 15 20:28:27 crc kubenswrapper[4735]: I0215 20:28:27.370789 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nczmn" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.236762 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln"] Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.238583 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.245543 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.268032 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln"] Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.302442 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt6vf\" (UniqueName: \"kubernetes.io/projected/880dc366-4d49-4093-b67d-715fea45d07a-kube-api-access-kt6vf\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.302509 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.302576 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.403422 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt6vf\" (UniqueName: \"kubernetes.io/projected/880dc366-4d49-4093-b67d-715fea45d07a-kube-api-access-kt6vf\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.403488 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.403546 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.404001 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.405103 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.425386 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt6vf\" (UniqueName: \"kubernetes.io/projected/880dc366-4d49-4093-b67d-715fea45d07a-kube-api-access-kt6vf\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.553910 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:33 crc kubenswrapper[4735]: I0215 20:28:33.762607 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln"] Feb 15 20:28:34 crc kubenswrapper[4735]: I0215 20:28:34.220531 4735 generic.go:334] "Generic (PLEG): container finished" podID="880dc366-4d49-4093-b67d-715fea45d07a" containerID="ae8a586d3312c8a1f3417d7078a507e161a2e5d83cc0068d51623e7977158ca9" exitCode=0 Feb 15 20:28:34 crc kubenswrapper[4735]: I0215 20:28:34.220735 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" event={"ID":"880dc366-4d49-4093-b67d-715fea45d07a","Type":"ContainerDied","Data":"ae8a586d3312c8a1f3417d7078a507e161a2e5d83cc0068d51623e7977158ca9"} Feb 15 20:28:34 crc kubenswrapper[4735]: I0215 20:28:34.221374 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" event={"ID":"880dc366-4d49-4093-b67d-715fea45d07a","Type":"ContainerStarted","Data":"3acaa4e8962d0ed30e4a8889481799469a289f1c14720f5490dce7368ad5ef5a"} Feb 15 20:28:36 crc kubenswrapper[4735]: I0215 20:28:36.235759 4735 generic.go:334] "Generic (PLEG): container finished" podID="880dc366-4d49-4093-b67d-715fea45d07a" containerID="fb837490fddabfc8a4d700c0900eb44f191ada53d8642bc71387245258f6f493" exitCode=0 Feb 15 20:28:36 crc kubenswrapper[4735]: I0215 20:28:36.235876 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" event={"ID":"880dc366-4d49-4093-b67d-715fea45d07a","Type":"ContainerDied","Data":"fb837490fddabfc8a4d700c0900eb44f191ada53d8642bc71387245258f6f493"} Feb 15 20:28:37 crc kubenswrapper[4735]: I0215 20:28:37.243278 4735 generic.go:334] "Generic (PLEG): container finished" podID="880dc366-4d49-4093-b67d-715fea45d07a" containerID="90d4f771decf87a50e8ae46e0ebc4d337fda1e4a4a3bfccb277b87e51ee77c52" exitCode=0 Feb 15 20:28:37 crc kubenswrapper[4735]: I0215 20:28:37.243334 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" event={"ID":"880dc366-4d49-4093-b67d-715fea45d07a","Type":"ContainerDied","Data":"90d4f771decf87a50e8ae46e0ebc4d337fda1e4a4a3bfccb277b87e51ee77c52"} Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.510008 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.694234 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-util\") pod \"880dc366-4d49-4093-b67d-715fea45d07a\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.694290 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-bundle\") pod \"880dc366-4d49-4093-b67d-715fea45d07a\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.694310 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt6vf\" (UniqueName: \"kubernetes.io/projected/880dc366-4d49-4093-b67d-715fea45d07a-kube-api-access-kt6vf\") pod \"880dc366-4d49-4093-b67d-715fea45d07a\" (UID: \"880dc366-4d49-4093-b67d-715fea45d07a\") " Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.695241 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-bundle" (OuterVolumeSpecName: "bundle") pod "880dc366-4d49-4093-b67d-715fea45d07a" (UID: "880dc366-4d49-4093-b67d-715fea45d07a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.702270 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880dc366-4d49-4093-b67d-715fea45d07a-kube-api-access-kt6vf" (OuterVolumeSpecName: "kube-api-access-kt6vf") pod "880dc366-4d49-4093-b67d-715fea45d07a" (UID: "880dc366-4d49-4093-b67d-715fea45d07a"). InnerVolumeSpecName "kube-api-access-kt6vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.723668 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-util" (OuterVolumeSpecName: "util") pod "880dc366-4d49-4093-b67d-715fea45d07a" (UID: "880dc366-4d49-4093-b67d-715fea45d07a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.795664 4735 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-util\") on node \"crc\" DevicePath \"\"" Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.795697 4735 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880dc366-4d49-4093-b67d-715fea45d07a-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:28:38 crc kubenswrapper[4735]: I0215 20:28:38.795707 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt6vf\" (UniqueName: \"kubernetes.io/projected/880dc366-4d49-4093-b67d-715fea45d07a-kube-api-access-kt6vf\") on node \"crc\" DevicePath \"\"" Feb 15 20:28:39 crc kubenswrapper[4735]: I0215 20:28:39.259411 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" event={"ID":"880dc366-4d49-4093-b67d-715fea45d07a","Type":"ContainerDied","Data":"3acaa4e8962d0ed30e4a8889481799469a289f1c14720f5490dce7368ad5ef5a"} Feb 15 20:28:39 crc kubenswrapper[4735]: I0215 20:28:39.259466 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3acaa4e8962d0ed30e4a8889481799469a289f1c14720f5490dce7368ad5ef5a" Feb 15 20:28:39 crc kubenswrapper[4735]: I0215 20:28:39.259510 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.444750 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-vc7vf"] Feb 15 20:28:40 crc kubenswrapper[4735]: E0215 20:28:40.445237 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="880dc366-4d49-4093-b67d-715fea45d07a" containerName="util" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.445253 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="880dc366-4d49-4093-b67d-715fea45d07a" containerName="util" Feb 15 20:28:40 crc kubenswrapper[4735]: E0215 20:28:40.445262 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="880dc366-4d49-4093-b67d-715fea45d07a" containerName="extract" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.445270 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="880dc366-4d49-4093-b67d-715fea45d07a" containerName="extract" Feb 15 20:28:40 crc kubenswrapper[4735]: E0215 20:28:40.445281 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="880dc366-4d49-4093-b67d-715fea45d07a" containerName="pull" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.445287 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="880dc366-4d49-4093-b67d-715fea45d07a" containerName="pull" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.445395 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="880dc366-4d49-4093-b67d-715fea45d07a" containerName="extract" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.445815 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-vc7vf" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.447683 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.448248 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.450635 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-8jw2c" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.464522 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-vc7vf"] Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.615747 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhngc\" (UniqueName: \"kubernetes.io/projected/16c72a4d-04eb-4a5f-af25-439d36f5c9de-kube-api-access-hhngc\") pod \"nmstate-operator-694c9596b7-vc7vf\" (UID: \"16c72a4d-04eb-4a5f-af25-439d36f5c9de\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-vc7vf" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.717483 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhngc\" (UniqueName: \"kubernetes.io/projected/16c72a4d-04eb-4a5f-af25-439d36f5c9de-kube-api-access-hhngc\") pod \"nmstate-operator-694c9596b7-vc7vf\" (UID: \"16c72a4d-04eb-4a5f-af25-439d36f5c9de\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-vc7vf" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.745492 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhngc\" (UniqueName: \"kubernetes.io/projected/16c72a4d-04eb-4a5f-af25-439d36f5c9de-kube-api-access-hhngc\") pod \"nmstate-operator-694c9596b7-vc7vf\" (UID: \"16c72a4d-04eb-4a5f-af25-439d36f5c9de\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-vc7vf" Feb 15 20:28:40 crc kubenswrapper[4735]: I0215 20:28:40.759894 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-vc7vf" Feb 15 20:28:41 crc kubenswrapper[4735]: I0215 20:28:41.178437 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-vc7vf"] Feb 15 20:28:41 crc kubenswrapper[4735]: I0215 20:28:41.268932 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-vc7vf" event={"ID":"16c72a4d-04eb-4a5f-af25-439d36f5c9de","Type":"ContainerStarted","Data":"030525a57c5eff22277df5841a45a5d5f64197fe8a119efc108704093729185b"} Feb 15 20:28:44 crc kubenswrapper[4735]: I0215 20:28:44.288189 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-vc7vf" event={"ID":"16c72a4d-04eb-4a5f-af25-439d36f5c9de","Type":"ContainerStarted","Data":"0192a09b90eadefa7c7e927474ab33bde48a7ea8d784267c443e5ee509be27ff"} Feb 15 20:28:44 crc kubenswrapper[4735]: I0215 20:28:44.302185 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-vc7vf" podStartSLOduration=1.973546889 podStartE2EDuration="4.302167861s" podCreationTimestamp="2026-02-15 20:28:40 +0000 UTC" firstStartedPulling="2026-02-15 20:28:41.189654375 +0000 UTC m=+729.055669998" lastFinishedPulling="2026-02-15 20:28:43.518275347 +0000 UTC m=+731.384290970" observedRunningTime="2026-02-15 20:28:44.301801981 +0000 UTC m=+732.167817624" watchObservedRunningTime="2026-02-15 20:28:44.302167861 +0000 UTC m=+732.168183484" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.304441 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.306178 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.307983 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-lj2bb" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.316166 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.319443 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.320144 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.323827 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.338420 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.350526 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-5grq7"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.351207 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.466641 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.467248 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470495 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-nmstate-lock\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470536 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-26pfp" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470566 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cecb1f2e-e07b-4ea1-924d-7802fa40fd6b-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-7gkdk\" (UID: \"cecb1f2e-e07b-4ea1-924d-7802fa40fd6b\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470623 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-ovs-socket\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470659 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7nlz\" (UniqueName: \"kubernetes.io/projected/6291dd59-7c75-4426-86d1-063a72855309-kube-api-access-f7nlz\") pod \"nmstate-metrics-58c85c668d-4gnfl\" (UID: \"6291dd59-7c75-4426-86d1-063a72855309\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470680 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-dbus-socket\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470724 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlsb7\" (UniqueName: \"kubernetes.io/projected/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-kube-api-access-nlsb7\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470737 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470768 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn8tb\" (UniqueName: \"kubernetes.io/projected/cecb1f2e-e07b-4ea1-924d-7802fa40fd6b-kube-api-access-zn8tb\") pod \"nmstate-webhook-866bcb46dc-7gkdk\" (UID: \"cecb1f2e-e07b-4ea1-924d-7802fa40fd6b\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.470843 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.479654 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571691 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn8tb\" (UniqueName: \"kubernetes.io/projected/cecb1f2e-e07b-4ea1-924d-7802fa40fd6b-kube-api-access-zn8tb\") pod \"nmstate-webhook-866bcb46dc-7gkdk\" (UID: \"cecb1f2e-e07b-4ea1-924d-7802fa40fd6b\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571745 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4d5d6279-ce2c-4382-900c-77563978d63a-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571769 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-nmstate-lock\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571798 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cecb1f2e-e07b-4ea1-924d-7802fa40fd6b-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-7gkdk\" (UID: \"cecb1f2e-e07b-4ea1-924d-7802fa40fd6b\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571828 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjmjh\" (UniqueName: \"kubernetes.io/projected/4d5d6279-ce2c-4382-900c-77563978d63a-kube-api-access-cjmjh\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571876 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-ovs-socket\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571898 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7nlz\" (UniqueName: \"kubernetes.io/projected/6291dd59-7c75-4426-86d1-063a72855309-kube-api-access-f7nlz\") pod \"nmstate-metrics-58c85c668d-4gnfl\" (UID: \"6291dd59-7c75-4426-86d1-063a72855309\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-dbus-socket\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571933 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlsb7\" (UniqueName: \"kubernetes.io/projected/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-kube-api-access-nlsb7\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.571964 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d5d6279-ce2c-4382-900c-77563978d63a-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.572046 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-nmstate-lock\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: E0215 20:28:45.572131 4735 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Feb 15 20:28:45 crc kubenswrapper[4735]: E0215 20:28:45.572174 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cecb1f2e-e07b-4ea1-924d-7802fa40fd6b-tls-key-pair podName:cecb1f2e-e07b-4ea1-924d-7802fa40fd6b nodeName:}" failed. No retries permitted until 2026-02-15 20:28:46.072159181 +0000 UTC m=+733.938174804 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/cecb1f2e-e07b-4ea1-924d-7802fa40fd6b-tls-key-pair") pod "nmstate-webhook-866bcb46dc-7gkdk" (UID: "cecb1f2e-e07b-4ea1-924d-7802fa40fd6b") : secret "openshift-nmstate-webhook" not found Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.572264 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-ovs-socket\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.572660 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-dbus-socket\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.595518 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlsb7\" (UniqueName: \"kubernetes.io/projected/ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e-kube-api-access-nlsb7\") pod \"nmstate-handler-5grq7\" (UID: \"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e\") " pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.596016 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn8tb\" (UniqueName: \"kubernetes.io/projected/cecb1f2e-e07b-4ea1-924d-7802fa40fd6b-kube-api-access-zn8tb\") pod \"nmstate-webhook-866bcb46dc-7gkdk\" (UID: \"cecb1f2e-e07b-4ea1-924d-7802fa40fd6b\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.609931 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7nlz\" (UniqueName: \"kubernetes.io/projected/6291dd59-7c75-4426-86d1-063a72855309-kube-api-access-f7nlz\") pod \"nmstate-metrics-58c85c668d-4gnfl\" (UID: \"6291dd59-7c75-4426-86d1-063a72855309\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.630600 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.644685 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-55bc4b8956-6szxr"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.645524 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.671423 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-55bc4b8956-6szxr"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.671633 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.673091 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d5d6279-ce2c-4382-900c-77563978d63a-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.675676 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4d5d6279-ce2c-4382-900c-77563978d63a-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.691124 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjmjh\" (UniqueName: \"kubernetes.io/projected/4d5d6279-ce2c-4382-900c-77563978d63a-kube-api-access-cjmjh\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:45 crc kubenswrapper[4735]: E0215 20:28:45.673875 4735 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 15 20:28:45 crc kubenswrapper[4735]: E0215 20:28:45.691291 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d5d6279-ce2c-4382-900c-77563978d63a-plugin-serving-cert podName:4d5d6279-ce2c-4382-900c-77563978d63a nodeName:}" failed. No retries permitted until 2026-02-15 20:28:46.191269651 +0000 UTC m=+734.057285274 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/4d5d6279-ce2c-4382-900c-77563978d63a-plugin-serving-cert") pod "nmstate-console-plugin-5c78fc5d65-6q5t8" (UID: "4d5d6279-ce2c-4382-900c-77563978d63a") : secret "plugin-serving-cert" not found Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.679771 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4d5d6279-ce2c-4382-900c-77563978d63a-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.717181 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjmjh\" (UniqueName: \"kubernetes.io/projected/4d5d6279-ce2c-4382-900c-77563978d63a-kube-api-access-cjmjh\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.797328 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-trusted-ca-bundle\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.797715 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-service-ca\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.797754 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-oauth-serving-cert\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.797794 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4mxb\" (UniqueName: \"kubernetes.io/projected/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-kube-api-access-g4mxb\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.797828 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-console-serving-cert\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.797843 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-console-oauth-config\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.797915 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-console-config\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.898901 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-service-ca\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.898969 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-oauth-serving-cert\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.898988 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4mxb\" (UniqueName: \"kubernetes.io/projected/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-kube-api-access-g4mxb\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.899023 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-console-serving-cert\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.899038 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-console-oauth-config\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.899066 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-console-config\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.899086 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-trusted-ca-bundle\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.900094 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-console-config\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.900113 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-trusted-ca-bundle\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.900216 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-service-ca\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.900396 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-oauth-serving-cert\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.902139 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-console-oauth-config\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.902272 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-console-serving-cert\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.909545 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl"] Feb 15 20:28:45 crc kubenswrapper[4735]: I0215 20:28:45.922720 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4mxb\" (UniqueName: \"kubernetes.io/projected/f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0-kube-api-access-g4mxb\") pod \"console-55bc4b8956-6szxr\" (UID: \"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0\") " pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.031906 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.100607 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cecb1f2e-e07b-4ea1-924d-7802fa40fd6b-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-7gkdk\" (UID: \"cecb1f2e-e07b-4ea1-924d-7802fa40fd6b\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.105077 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cecb1f2e-e07b-4ea1-924d-7802fa40fd6b-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-7gkdk\" (UID: \"cecb1f2e-e07b-4ea1-924d-7802fa40fd6b\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.201805 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d5d6279-ce2c-4382-900c-77563978d63a-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.207617 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d5d6279-ce2c-4382-900c-77563978d63a-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-6q5t8\" (UID: \"4d5d6279-ce2c-4382-900c-77563978d63a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.237017 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.246908 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-55bc4b8956-6szxr"] Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.301329 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl" event={"ID":"6291dd59-7c75-4426-86d1-063a72855309","Type":"ContainerStarted","Data":"a8dac5f20c4e7746824a2a10fadb94d0534e82d3625c3f146016919f1ef62467"} Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.302551 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-55bc4b8956-6szxr" event={"ID":"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0","Type":"ContainerStarted","Data":"74964fd96b4a25c69dcc520c77b98318de693568a1dfae3386bf5f975d0b41ac"} Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.305852 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5grq7" event={"ID":"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e","Type":"ContainerStarted","Data":"a5de1087703a2ad5da8356cc24912db92a04f253fff938d77103cfbc792c3026"} Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.393839 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.580060 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8"] Feb 15 20:28:46 crc kubenswrapper[4735]: W0215 20:28:46.587165 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d5d6279_ce2c_4382_900c_77563978d63a.slice/crio-968a61e560a86b6c8ae0eb93bf7123eb6c9272edf253c679e8f04818c777a139 WatchSource:0}: Error finding container 968a61e560a86b6c8ae0eb93bf7123eb6c9272edf253c679e8f04818c777a139: Status 404 returned error can't find the container with id 968a61e560a86b6c8ae0eb93bf7123eb6c9272edf253c679e8f04818c777a139 Feb 15 20:28:46 crc kubenswrapper[4735]: I0215 20:28:46.671490 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk"] Feb 15 20:28:46 crc kubenswrapper[4735]: W0215 20:28:46.684119 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcecb1f2e_e07b_4ea1_924d_7802fa40fd6b.slice/crio-0a4fad4aa338b703aee74557cbac73dbc5619303ed733e10ed3a0279aa4a2b42 WatchSource:0}: Error finding container 0a4fad4aa338b703aee74557cbac73dbc5619303ed733e10ed3a0279aa4a2b42: Status 404 returned error can't find the container with id 0a4fad4aa338b703aee74557cbac73dbc5619303ed733e10ed3a0279aa4a2b42 Feb 15 20:28:47 crc kubenswrapper[4735]: I0215 20:28:47.311849 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-55bc4b8956-6szxr" event={"ID":"f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0","Type":"ContainerStarted","Data":"04b954f5a6fb0cbdf5026ea90a67d5db24d64796f1c79befb34df4a1feaa8f01"} Feb 15 20:28:47 crc kubenswrapper[4735]: I0215 20:28:47.314559 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" event={"ID":"cecb1f2e-e07b-4ea1-924d-7802fa40fd6b","Type":"ContainerStarted","Data":"0a4fad4aa338b703aee74557cbac73dbc5619303ed733e10ed3a0279aa4a2b42"} Feb 15 20:28:47 crc kubenswrapper[4735]: I0215 20:28:47.315197 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" event={"ID":"4d5d6279-ce2c-4382-900c-77563978d63a","Type":"ContainerStarted","Data":"968a61e560a86b6c8ae0eb93bf7123eb6c9272edf253c679e8f04818c777a139"} Feb 15 20:28:47 crc kubenswrapper[4735]: I0215 20:28:47.332872 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-55bc4b8956-6szxr" podStartSLOduration=2.332857864 podStartE2EDuration="2.332857864s" podCreationTimestamp="2026-02-15 20:28:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:28:47.32921705 +0000 UTC m=+735.195232673" watchObservedRunningTime="2026-02-15 20:28:47.332857864 +0000 UTC m=+735.198873487" Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.330318 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" event={"ID":"4d5d6279-ce2c-4382-900c-77563978d63a","Type":"ContainerStarted","Data":"9edb385d083acafd6d8a04095803d9c99851af76231b30f4ad817c3ff9a58ce0"} Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.333765 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5grq7" event={"ID":"ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e","Type":"ContainerStarted","Data":"2e020e8a37ca5318173353dddd00390c6d778c3451d94d91e62c0d6d37e8e290"} Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.334053 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.337317 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" event={"ID":"cecb1f2e-e07b-4ea1-924d-7802fa40fd6b","Type":"ContainerStarted","Data":"a33c6085d988f6ca7c02f0f1ab72f217714fdd88bccca5c36a924de752f41b9f"} Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.338084 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.340179 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl" event={"ID":"6291dd59-7c75-4426-86d1-063a72855309","Type":"ContainerStarted","Data":"26a2efe0ec1ba78ac2ca600cd6bae570a2e8d3e9ac6326a9f557badbd046c739"} Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.357308 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-6q5t8" podStartSLOduration=2.296475931 podStartE2EDuration="4.357288913s" podCreationTimestamp="2026-02-15 20:28:45 +0000 UTC" firstStartedPulling="2026-02-15 20:28:46.589423141 +0000 UTC m=+734.455438764" lastFinishedPulling="2026-02-15 20:28:48.650236123 +0000 UTC m=+736.516251746" observedRunningTime="2026-02-15 20:28:49.352387526 +0000 UTC m=+737.218403139" watchObservedRunningTime="2026-02-15 20:28:49.357288913 +0000 UTC m=+737.223304536" Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.378296 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-5grq7" podStartSLOduration=1.450149636 podStartE2EDuration="4.378271187s" podCreationTimestamp="2026-02-15 20:28:45 +0000 UTC" firstStartedPulling="2026-02-15 20:28:45.71896045 +0000 UTC m=+733.584976073" lastFinishedPulling="2026-02-15 20:28:48.647081961 +0000 UTC m=+736.513097624" observedRunningTime="2026-02-15 20:28:49.373992237 +0000 UTC m=+737.240007850" watchObservedRunningTime="2026-02-15 20:28:49.378271187 +0000 UTC m=+737.244286850" Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.404584 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" podStartSLOduration=2.441918064 podStartE2EDuration="4.40456281s" podCreationTimestamp="2026-02-15 20:28:45 +0000 UTC" firstStartedPulling="2026-02-15 20:28:46.686846568 +0000 UTC m=+734.552862211" lastFinishedPulling="2026-02-15 20:28:48.649491334 +0000 UTC m=+736.515506957" observedRunningTime="2026-02-15 20:28:49.404450357 +0000 UTC m=+737.270466010" watchObservedRunningTime="2026-02-15 20:28:49.40456281 +0000 UTC m=+737.270578463" Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.680371 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:28:49 crc kubenswrapper[4735]: I0215 20:28:49.680742 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:28:51 crc kubenswrapper[4735]: I0215 20:28:51.352817 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl" event={"ID":"6291dd59-7c75-4426-86d1-063a72855309","Type":"ContainerStarted","Data":"1828af2911cf934429e56c8241655697df45ae14f786dec646345ff3bbc9ce18"} Feb 15 20:28:51 crc kubenswrapper[4735]: I0215 20:28:51.373598 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-4gnfl" podStartSLOduration=1.4349353009999999 podStartE2EDuration="6.37358309s" podCreationTimestamp="2026-02-15 20:28:45 +0000 UTC" firstStartedPulling="2026-02-15 20:28:45.915605743 +0000 UTC m=+733.781621366" lastFinishedPulling="2026-02-15 20:28:50.854253532 +0000 UTC m=+738.720269155" observedRunningTime="2026-02-15 20:28:51.370360196 +0000 UTC m=+739.236375819" watchObservedRunningTime="2026-02-15 20:28:51.37358309 +0000 UTC m=+739.239598713" Feb 15 20:28:55 crc kubenswrapper[4735]: I0215 20:28:55.698523 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-5grq7" Feb 15 20:28:56 crc kubenswrapper[4735]: I0215 20:28:56.033559 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:56 crc kubenswrapper[4735]: I0215 20:28:56.034065 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:56 crc kubenswrapper[4735]: I0215 20:28:56.039835 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:56 crc kubenswrapper[4735]: I0215 20:28:56.387874 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-55bc4b8956-6szxr" Feb 15 20:28:56 crc kubenswrapper[4735]: I0215 20:28:56.449973 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k64kb"] Feb 15 20:29:06 crc kubenswrapper[4735]: I0215 20:29:06.244832 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" Feb 15 20:29:12 crc kubenswrapper[4735]: I0215 20:29:12.187914 4735 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 15 20:29:19 crc kubenswrapper[4735]: I0215 20:29:19.680195 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:29:19 crc kubenswrapper[4735]: I0215 20:29:19.680706 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.381311 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc"] Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.383582 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.386360 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.404868 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc"] Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.467130 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.467186 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw8vp\" (UniqueName: \"kubernetes.io/projected/9afae777-8644-4e09-8166-55b3e1e34de9-kube-api-access-jw8vp\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.467202 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.506463 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-k64kb" podUID="e74e438d-61ef-403a-b958-8704f195f0ce" containerName="console" containerID="cri-o://9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853" gracePeriod=15 Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.568920 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.569021 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw8vp\" (UniqueName: \"kubernetes.io/projected/9afae777-8644-4e09-8166-55b3e1e34de9-kube-api-access-jw8vp\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.569073 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.569491 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.569599 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.587974 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw8vp\" (UniqueName: \"kubernetes.io/projected/9afae777-8644-4e09-8166-55b3e1e34de9-kube-api-access-jw8vp\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.710247 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.863456 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k64kb_e74e438d-61ef-403a-b958-8704f195f0ce/console/0.log" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.863756 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.917144 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc"] Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.956018 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k64kb_e74e438d-61ef-403a-b958-8704f195f0ce/console/0.log" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.956273 4735 generic.go:334] "Generic (PLEG): container finished" podID="e74e438d-61ef-403a-b958-8704f195f0ce" containerID="9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853" exitCode=2 Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.956372 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k64kb" event={"ID":"e74e438d-61ef-403a-b958-8704f195f0ce","Type":"ContainerDied","Data":"9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853"} Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.956454 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k64kb" event={"ID":"e74e438d-61ef-403a-b958-8704f195f0ce","Type":"ContainerDied","Data":"921be920b01c813ad90e0ba7b7ace4b4ccb8c46aed15fb8de3061bb4f4de21e0"} Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.956459 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k64kb" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.956474 4735 scope.go:117] "RemoveContainer" containerID="9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.960654 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" event={"ID":"9afae777-8644-4e09-8166-55b3e1e34de9","Type":"ContainerStarted","Data":"e7acdf62a01102a908913df5a459186260132ba7b226125404dbca1bade6b76d"} Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.973199 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvkzc\" (UniqueName: \"kubernetes.io/projected/e74e438d-61ef-403a-b958-8704f195f0ce-kube-api-access-tvkzc\") pod \"e74e438d-61ef-403a-b958-8704f195f0ce\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.973242 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-console-config\") pod \"e74e438d-61ef-403a-b958-8704f195f0ce\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.973289 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-oauth-config\") pod \"e74e438d-61ef-403a-b958-8704f195f0ce\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.973315 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-oauth-serving-cert\") pod \"e74e438d-61ef-403a-b958-8704f195f0ce\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.973336 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-service-ca\") pod \"e74e438d-61ef-403a-b958-8704f195f0ce\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.973384 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-trusted-ca-bundle\") pod \"e74e438d-61ef-403a-b958-8704f195f0ce\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.973449 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-serving-cert\") pod \"e74e438d-61ef-403a-b958-8704f195f0ce\" (UID: \"e74e438d-61ef-403a-b958-8704f195f0ce\") " Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.974071 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e74e438d-61ef-403a-b958-8704f195f0ce" (UID: "e74e438d-61ef-403a-b958-8704f195f0ce"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.974095 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e74e438d-61ef-403a-b958-8704f195f0ce" (UID: "e74e438d-61ef-403a-b958-8704f195f0ce"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.974294 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-console-config" (OuterVolumeSpecName: "console-config") pod "e74e438d-61ef-403a-b958-8704f195f0ce" (UID: "e74e438d-61ef-403a-b958-8704f195f0ce"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.974390 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-service-ca" (OuterVolumeSpecName: "service-ca") pod "e74e438d-61ef-403a-b958-8704f195f0ce" (UID: "e74e438d-61ef-403a-b958-8704f195f0ce"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.975330 4735 scope.go:117] "RemoveContainer" containerID="9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853" Feb 15 20:29:21 crc kubenswrapper[4735]: E0215 20:29:21.975719 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853\": container with ID starting with 9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853 not found: ID does not exist" containerID="9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.975791 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853"} err="failed to get container status \"9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853\": rpc error: code = NotFound desc = could not find container \"9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853\": container with ID starting with 9e8e4d8a8dd35f864b0a6b153623d8b83f14feb1bb831d5ee4651737daa9f853 not found: ID does not exist" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.976964 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e74e438d-61ef-403a-b958-8704f195f0ce" (UID: "e74e438d-61ef-403a-b958-8704f195f0ce"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.977646 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e74e438d-61ef-403a-b958-8704f195f0ce-kube-api-access-tvkzc" (OuterVolumeSpecName: "kube-api-access-tvkzc") pod "e74e438d-61ef-403a-b958-8704f195f0ce" (UID: "e74e438d-61ef-403a-b958-8704f195f0ce"). InnerVolumeSpecName "kube-api-access-tvkzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:29:21 crc kubenswrapper[4735]: I0215 20:29:21.977648 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e74e438d-61ef-403a-b958-8704f195f0ce" (UID: "e74e438d-61ef-403a-b958-8704f195f0ce"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.075092 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvkzc\" (UniqueName: \"kubernetes.io/projected/e74e438d-61ef-403a-b958-8704f195f0ce-kube-api-access-tvkzc\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.075129 4735 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-console-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.075139 4735 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.075148 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-service-ca\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.075156 4735 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.075164 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e74e438d-61ef-403a-b958-8704f195f0ce-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.075171 4735 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e74e438d-61ef-403a-b958-8704f195f0ce-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.283187 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k64kb"] Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.286225 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-k64kb"] Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.896603 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e74e438d-61ef-403a-b958-8704f195f0ce" path="/var/lib/kubelet/pods/e74e438d-61ef-403a-b958-8704f195f0ce/volumes" Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.973535 4735 generic.go:334] "Generic (PLEG): container finished" podID="9afae777-8644-4e09-8166-55b3e1e34de9" containerID="6a5193834bf01d7f7c7747a4a69568a982aa1870699b4d3d1ee278e03487eb31" exitCode=0 Feb 15 20:29:22 crc kubenswrapper[4735]: I0215 20:29:22.973593 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" event={"ID":"9afae777-8644-4e09-8166-55b3e1e34de9","Type":"ContainerDied","Data":"6a5193834bf01d7f7c7747a4a69568a982aa1870699b4d3d1ee278e03487eb31"} Feb 15 20:29:23 crc kubenswrapper[4735]: I0215 20:29:23.730252 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rrp27"] Feb 15 20:29:23 crc kubenswrapper[4735]: E0215 20:29:23.730446 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74e438d-61ef-403a-b958-8704f195f0ce" containerName="console" Feb 15 20:29:23 crc kubenswrapper[4735]: I0215 20:29:23.730457 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74e438d-61ef-403a-b958-8704f195f0ce" containerName="console" Feb 15 20:29:23 crc kubenswrapper[4735]: I0215 20:29:23.730576 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e74e438d-61ef-403a-b958-8704f195f0ce" containerName="console" Feb 15 20:29:23 crc kubenswrapper[4735]: I0215 20:29:23.731247 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:23 crc kubenswrapper[4735]: I0215 20:29:23.743554 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rrp27"] Feb 15 20:29:23 crc kubenswrapper[4735]: I0215 20:29:23.901140 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-utilities\") pod \"redhat-operators-rrp27\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:23 crc kubenswrapper[4735]: I0215 20:29:23.901226 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpwwj\" (UniqueName: \"kubernetes.io/projected/eba24811-9f0a-486b-b2d2-977950b95d61-kube-api-access-cpwwj\") pod \"redhat-operators-rrp27\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:23 crc kubenswrapper[4735]: I0215 20:29:23.901248 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-catalog-content\") pod \"redhat-operators-rrp27\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.002063 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpwwj\" (UniqueName: \"kubernetes.io/projected/eba24811-9f0a-486b-b2d2-977950b95d61-kube-api-access-cpwwj\") pod \"redhat-operators-rrp27\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.002104 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-catalog-content\") pod \"redhat-operators-rrp27\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.002258 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-utilities\") pod \"redhat-operators-rrp27\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.002690 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-utilities\") pod \"redhat-operators-rrp27\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.003232 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-catalog-content\") pod \"redhat-operators-rrp27\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.029505 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpwwj\" (UniqueName: \"kubernetes.io/projected/eba24811-9f0a-486b-b2d2-977950b95d61-kube-api-access-cpwwj\") pod \"redhat-operators-rrp27\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.055257 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.275156 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rrp27"] Feb 15 20:29:24 crc kubenswrapper[4735]: W0215 20:29:24.301916 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeba24811_9f0a_486b_b2d2_977950b95d61.slice/crio-26d6624a7ea31ce1bed1e68058bfa41e0415e688dc0f91fe1914b1d4cb0fc471 WatchSource:0}: Error finding container 26d6624a7ea31ce1bed1e68058bfa41e0415e688dc0f91fe1914b1d4cb0fc471: Status 404 returned error can't find the container with id 26d6624a7ea31ce1bed1e68058bfa41e0415e688dc0f91fe1914b1d4cb0fc471 Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.982793 4735 generic.go:334] "Generic (PLEG): container finished" podID="eba24811-9f0a-486b-b2d2-977950b95d61" containerID="6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a" exitCode=0 Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.982898 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrp27" event={"ID":"eba24811-9f0a-486b-b2d2-977950b95d61","Type":"ContainerDied","Data":"6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a"} Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.983116 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrp27" event={"ID":"eba24811-9f0a-486b-b2d2-977950b95d61","Type":"ContainerStarted","Data":"26d6624a7ea31ce1bed1e68058bfa41e0415e688dc0f91fe1914b1d4cb0fc471"} Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.985531 4735 generic.go:334] "Generic (PLEG): container finished" podID="9afae777-8644-4e09-8166-55b3e1e34de9" containerID="c8e075ce2ea94f3b10ba0a766c5dea51ac33c12c0fdb161895768174ef3f187a" exitCode=0 Feb 15 20:29:24 crc kubenswrapper[4735]: I0215 20:29:24.985561 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" event={"ID":"9afae777-8644-4e09-8166-55b3e1e34de9","Type":"ContainerDied","Data":"c8e075ce2ea94f3b10ba0a766c5dea51ac33c12c0fdb161895768174ef3f187a"} Feb 15 20:29:25 crc kubenswrapper[4735]: I0215 20:29:25.997177 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrp27" event={"ID":"eba24811-9f0a-486b-b2d2-977950b95d61","Type":"ContainerStarted","Data":"f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988"} Feb 15 20:29:26 crc kubenswrapper[4735]: I0215 20:29:26.000467 4735 generic.go:334] "Generic (PLEG): container finished" podID="9afae777-8644-4e09-8166-55b3e1e34de9" containerID="f6fc189756cdf044899a1e07b100eedfa9f92e592b2c46093164ae1c6881e80b" exitCode=0 Feb 15 20:29:26 crc kubenswrapper[4735]: I0215 20:29:26.000527 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" event={"ID":"9afae777-8644-4e09-8166-55b3e1e34de9","Type":"ContainerDied","Data":"f6fc189756cdf044899a1e07b100eedfa9f92e592b2c46093164ae1c6881e80b"} Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.008726 4735 generic.go:334] "Generic (PLEG): container finished" podID="eba24811-9f0a-486b-b2d2-977950b95d61" containerID="f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988" exitCode=0 Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.008788 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrp27" event={"ID":"eba24811-9f0a-486b-b2d2-977950b95d61","Type":"ContainerDied","Data":"f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988"} Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.330109 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.444150 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-util\") pod \"9afae777-8644-4e09-8166-55b3e1e34de9\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.444252 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw8vp\" (UniqueName: \"kubernetes.io/projected/9afae777-8644-4e09-8166-55b3e1e34de9-kube-api-access-jw8vp\") pod \"9afae777-8644-4e09-8166-55b3e1e34de9\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.444345 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-bundle\") pod \"9afae777-8644-4e09-8166-55b3e1e34de9\" (UID: \"9afae777-8644-4e09-8166-55b3e1e34de9\") " Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.445243 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-bundle" (OuterVolumeSpecName: "bundle") pod "9afae777-8644-4e09-8166-55b3e1e34de9" (UID: "9afae777-8644-4e09-8166-55b3e1e34de9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.453449 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9afae777-8644-4e09-8166-55b3e1e34de9-kube-api-access-jw8vp" (OuterVolumeSpecName: "kube-api-access-jw8vp") pod "9afae777-8644-4e09-8166-55b3e1e34de9" (UID: "9afae777-8644-4e09-8166-55b3e1e34de9"). InnerVolumeSpecName "kube-api-access-jw8vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.508529 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-util" (OuterVolumeSpecName: "util") pod "9afae777-8644-4e09-8166-55b3e1e34de9" (UID: "9afae777-8644-4e09-8166-55b3e1e34de9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.545232 4735 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.545270 4735 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9afae777-8644-4e09-8166-55b3e1e34de9-util\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:27 crc kubenswrapper[4735]: I0215 20:29:27.545283 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw8vp\" (UniqueName: \"kubernetes.io/projected/9afae777-8644-4e09-8166-55b3e1e34de9-kube-api-access-jw8vp\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:28 crc kubenswrapper[4735]: I0215 20:29:28.026161 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrp27" event={"ID":"eba24811-9f0a-486b-b2d2-977950b95d61","Type":"ContainerStarted","Data":"2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885"} Feb 15 20:29:28 crc kubenswrapper[4735]: I0215 20:29:28.033901 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" Feb 15 20:29:28 crc kubenswrapper[4735]: I0215 20:29:28.039032 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc" event={"ID":"9afae777-8644-4e09-8166-55b3e1e34de9","Type":"ContainerDied","Data":"e7acdf62a01102a908913df5a459186260132ba7b226125404dbca1bade6b76d"} Feb 15 20:29:28 crc kubenswrapper[4735]: I0215 20:29:28.039077 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7acdf62a01102a908913df5a459186260132ba7b226125404dbca1bade6b76d" Feb 15 20:29:28 crc kubenswrapper[4735]: I0215 20:29:28.071864 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rrp27" podStartSLOduration=2.524017227 podStartE2EDuration="5.071840468s" podCreationTimestamp="2026-02-15 20:29:23 +0000 UTC" firstStartedPulling="2026-02-15 20:29:24.983831978 +0000 UTC m=+772.849847601" lastFinishedPulling="2026-02-15 20:29:27.531655209 +0000 UTC m=+775.397670842" observedRunningTime="2026-02-15 20:29:28.04993764 +0000 UTC m=+775.915953283" watchObservedRunningTime="2026-02-15 20:29:28.071840468 +0000 UTC m=+775.937856091" Feb 15 20:29:34 crc kubenswrapper[4735]: I0215 20:29:34.056000 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:34 crc kubenswrapper[4735]: I0215 20:29:34.056587 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:35 crc kubenswrapper[4735]: I0215 20:29:35.106208 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rrp27" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" containerName="registry-server" probeResult="failure" output=< Feb 15 20:29:35 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:29:35 crc kubenswrapper[4735]: > Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.796324 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7"] Feb 15 20:29:36 crc kubenswrapper[4735]: E0215 20:29:36.796739 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9afae777-8644-4e09-8166-55b3e1e34de9" containerName="extract" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.796750 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9afae777-8644-4e09-8166-55b3e1e34de9" containerName="extract" Feb 15 20:29:36 crc kubenswrapper[4735]: E0215 20:29:36.796765 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9afae777-8644-4e09-8166-55b3e1e34de9" containerName="util" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.796770 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9afae777-8644-4e09-8166-55b3e1e34de9" containerName="util" Feb 15 20:29:36 crc kubenswrapper[4735]: E0215 20:29:36.796779 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9afae777-8644-4e09-8166-55b3e1e34de9" containerName="pull" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.796785 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9afae777-8644-4e09-8166-55b3e1e34de9" containerName="pull" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.796887 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9afae777-8644-4e09-8166-55b3e1e34de9" containerName="extract" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.797215 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.804810 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.804818 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.805083 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.805119 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-qx6wx" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.805241 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.826538 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7"] Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.967174 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c27da055-adee-4524-93ce-e2e83cf5966a-webhook-cert\") pod \"metallb-operator-controller-manager-79df65bf49-hzgd7\" (UID: \"c27da055-adee-4524-93ce-e2e83cf5966a\") " pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.967226 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzb7x\" (UniqueName: \"kubernetes.io/projected/c27da055-adee-4524-93ce-e2e83cf5966a-kube-api-access-gzb7x\") pod \"metallb-operator-controller-manager-79df65bf49-hzgd7\" (UID: \"c27da055-adee-4524-93ce-e2e83cf5966a\") " pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:36 crc kubenswrapper[4735]: I0215 20:29:36.967340 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c27da055-adee-4524-93ce-e2e83cf5966a-apiservice-cert\") pod \"metallb-operator-controller-manager-79df65bf49-hzgd7\" (UID: \"c27da055-adee-4524-93ce-e2e83cf5966a\") " pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.069639 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c27da055-adee-4524-93ce-e2e83cf5966a-webhook-cert\") pod \"metallb-operator-controller-manager-79df65bf49-hzgd7\" (UID: \"c27da055-adee-4524-93ce-e2e83cf5966a\") " pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.069682 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzb7x\" (UniqueName: \"kubernetes.io/projected/c27da055-adee-4524-93ce-e2e83cf5966a-kube-api-access-gzb7x\") pod \"metallb-operator-controller-manager-79df65bf49-hzgd7\" (UID: \"c27da055-adee-4524-93ce-e2e83cf5966a\") " pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.069733 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c27da055-adee-4524-93ce-e2e83cf5966a-apiservice-cert\") pod \"metallb-operator-controller-manager-79df65bf49-hzgd7\" (UID: \"c27da055-adee-4524-93ce-e2e83cf5966a\") " pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.075642 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c27da055-adee-4524-93ce-e2e83cf5966a-apiservice-cert\") pod \"metallb-operator-controller-manager-79df65bf49-hzgd7\" (UID: \"c27da055-adee-4524-93ce-e2e83cf5966a\") " pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.078698 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c27da055-adee-4524-93ce-e2e83cf5966a-webhook-cert\") pod \"metallb-operator-controller-manager-79df65bf49-hzgd7\" (UID: \"c27da055-adee-4524-93ce-e2e83cf5966a\") " pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.085719 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzb7x\" (UniqueName: \"kubernetes.io/projected/c27da055-adee-4524-93ce-e2e83cf5966a-kube-api-access-gzb7x\") pod \"metallb-operator-controller-manager-79df65bf49-hzgd7\" (UID: \"c27da055-adee-4524-93ce-e2e83cf5966a\") " pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.111654 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.135312 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm"] Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.136108 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.144369 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.144435 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-hktdn" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.144671 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.148509 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm"] Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.280688 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqlwv\" (UniqueName: \"kubernetes.io/projected/3516fdcd-aa1a-48ea-9f44-b928ca524e28-kube-api-access-mqlwv\") pod \"metallb-operator-webhook-server-7f47cf9495-7nwhm\" (UID: \"3516fdcd-aa1a-48ea-9f44-b928ca524e28\") " pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.280966 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3516fdcd-aa1a-48ea-9f44-b928ca524e28-apiservice-cert\") pod \"metallb-operator-webhook-server-7f47cf9495-7nwhm\" (UID: \"3516fdcd-aa1a-48ea-9f44-b928ca524e28\") " pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.280999 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3516fdcd-aa1a-48ea-9f44-b928ca524e28-webhook-cert\") pod \"metallb-operator-webhook-server-7f47cf9495-7nwhm\" (UID: \"3516fdcd-aa1a-48ea-9f44-b928ca524e28\") " pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.382185 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqlwv\" (UniqueName: \"kubernetes.io/projected/3516fdcd-aa1a-48ea-9f44-b928ca524e28-kube-api-access-mqlwv\") pod \"metallb-operator-webhook-server-7f47cf9495-7nwhm\" (UID: \"3516fdcd-aa1a-48ea-9f44-b928ca524e28\") " pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.382258 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3516fdcd-aa1a-48ea-9f44-b928ca524e28-apiservice-cert\") pod \"metallb-operator-webhook-server-7f47cf9495-7nwhm\" (UID: \"3516fdcd-aa1a-48ea-9f44-b928ca524e28\") " pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.382300 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3516fdcd-aa1a-48ea-9f44-b928ca524e28-webhook-cert\") pod \"metallb-operator-webhook-server-7f47cf9495-7nwhm\" (UID: \"3516fdcd-aa1a-48ea-9f44-b928ca524e28\") " pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.395597 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3516fdcd-aa1a-48ea-9f44-b928ca524e28-webhook-cert\") pod \"metallb-operator-webhook-server-7f47cf9495-7nwhm\" (UID: \"3516fdcd-aa1a-48ea-9f44-b928ca524e28\") " pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.403164 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3516fdcd-aa1a-48ea-9f44-b928ca524e28-apiservice-cert\") pod \"metallb-operator-webhook-server-7f47cf9495-7nwhm\" (UID: \"3516fdcd-aa1a-48ea-9f44-b928ca524e28\") " pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.423441 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqlwv\" (UniqueName: \"kubernetes.io/projected/3516fdcd-aa1a-48ea-9f44-b928ca524e28-kube-api-access-mqlwv\") pod \"metallb-operator-webhook-server-7f47cf9495-7nwhm\" (UID: \"3516fdcd-aa1a-48ea-9f44-b928ca524e28\") " pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.472603 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7"] Feb 15 20:29:37 crc kubenswrapper[4735]: W0215 20:29:37.480603 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc27da055_adee_4524_93ce_e2e83cf5966a.slice/crio-079c03c841029a30de860776d80c9cc72ebb5d2eccd23c578937708ea186d0c1 WatchSource:0}: Error finding container 079c03c841029a30de860776d80c9cc72ebb5d2eccd23c578937708ea186d0c1: Status 404 returned error can't find the container with id 079c03c841029a30de860776d80c9cc72ebb5d2eccd23c578937708ea186d0c1 Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.480932 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:37 crc kubenswrapper[4735]: I0215 20:29:37.692120 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm"] Feb 15 20:29:37 crc kubenswrapper[4735]: W0215 20:29:37.705731 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3516fdcd_aa1a_48ea_9f44_b928ca524e28.slice/crio-54804a9817ea359797f02b573f50dab4b2713c77e3194fd6510036fc07c6ed0a WatchSource:0}: Error finding container 54804a9817ea359797f02b573f50dab4b2713c77e3194fd6510036fc07c6ed0a: Status 404 returned error can't find the container with id 54804a9817ea359797f02b573f50dab4b2713c77e3194fd6510036fc07c6ed0a Feb 15 20:29:38 crc kubenswrapper[4735]: I0215 20:29:38.092710 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" event={"ID":"c27da055-adee-4524-93ce-e2e83cf5966a","Type":"ContainerStarted","Data":"079c03c841029a30de860776d80c9cc72ebb5d2eccd23c578937708ea186d0c1"} Feb 15 20:29:38 crc kubenswrapper[4735]: I0215 20:29:38.093620 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" event={"ID":"3516fdcd-aa1a-48ea-9f44-b928ca524e28","Type":"ContainerStarted","Data":"54804a9817ea359797f02b573f50dab4b2713c77e3194fd6510036fc07c6ed0a"} Feb 15 20:29:44 crc kubenswrapper[4735]: I0215 20:29:44.113197 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:44 crc kubenswrapper[4735]: I0215 20:29:44.152567 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:45 crc kubenswrapper[4735]: I0215 20:29:45.137842 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" event={"ID":"3516fdcd-aa1a-48ea-9f44-b928ca524e28","Type":"ContainerStarted","Data":"a41edee0f2b96f1e00b83bacb74bccac53257ae8ba62dedc42c5a9b181fd0291"} Feb 15 20:29:45 crc kubenswrapper[4735]: I0215 20:29:45.138189 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:29:45 crc kubenswrapper[4735]: I0215 20:29:45.139712 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" event={"ID":"c27da055-adee-4524-93ce-e2e83cf5966a","Type":"ContainerStarted","Data":"7d67a0243eece6862a2db22a95ca63340996cad5051fd5a1f90f370f1657eb93"} Feb 15 20:29:45 crc kubenswrapper[4735]: I0215 20:29:45.139862 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:29:45 crc kubenswrapper[4735]: I0215 20:29:45.161093 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" podStartSLOduration=1.334063429 podStartE2EDuration="8.161076486s" podCreationTimestamp="2026-02-15 20:29:37 +0000 UTC" firstStartedPulling="2026-02-15 20:29:37.709311496 +0000 UTC m=+785.575327119" lastFinishedPulling="2026-02-15 20:29:44.536324553 +0000 UTC m=+792.402340176" observedRunningTime="2026-02-15 20:29:45.157957595 +0000 UTC m=+793.023973218" watchObservedRunningTime="2026-02-15 20:29:45.161076486 +0000 UTC m=+793.027092109" Feb 15 20:29:45 crc kubenswrapper[4735]: I0215 20:29:45.182766 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" podStartSLOduration=2.149478037 podStartE2EDuration="9.182747692s" podCreationTimestamp="2026-02-15 20:29:36 +0000 UTC" firstStartedPulling="2026-02-15 20:29:37.487884311 +0000 UTC m=+785.353899934" lastFinishedPulling="2026-02-15 20:29:44.521153966 +0000 UTC m=+792.387169589" observedRunningTime="2026-02-15 20:29:45.17885113 +0000 UTC m=+793.044866753" watchObservedRunningTime="2026-02-15 20:29:45.182747692 +0000 UTC m=+793.048763325" Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.325218 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rrp27"] Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.325483 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rrp27" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" containerName="registry-server" containerID="cri-o://2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885" gracePeriod=2 Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.719109 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.804052 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-catalog-content\") pod \"eba24811-9f0a-486b-b2d2-977950b95d61\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.804115 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpwwj\" (UniqueName: \"kubernetes.io/projected/eba24811-9f0a-486b-b2d2-977950b95d61-kube-api-access-cpwwj\") pod \"eba24811-9f0a-486b-b2d2-977950b95d61\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.804216 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-utilities\") pod \"eba24811-9f0a-486b-b2d2-977950b95d61\" (UID: \"eba24811-9f0a-486b-b2d2-977950b95d61\") " Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.805019 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-utilities" (OuterVolumeSpecName: "utilities") pod "eba24811-9f0a-486b-b2d2-977950b95d61" (UID: "eba24811-9f0a-486b-b2d2-977950b95d61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.809599 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eba24811-9f0a-486b-b2d2-977950b95d61-kube-api-access-cpwwj" (OuterVolumeSpecName: "kube-api-access-cpwwj") pod "eba24811-9f0a-486b-b2d2-977950b95d61" (UID: "eba24811-9f0a-486b-b2d2-977950b95d61"). InnerVolumeSpecName "kube-api-access-cpwwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.905237 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpwwj\" (UniqueName: \"kubernetes.io/projected/eba24811-9f0a-486b-b2d2-977950b95d61-kube-api-access-cpwwj\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.905264 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:46 crc kubenswrapper[4735]: I0215 20:29:46.915157 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eba24811-9f0a-486b-b2d2-977950b95d61" (UID: "eba24811-9f0a-486b-b2d2-977950b95d61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.006435 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eba24811-9f0a-486b-b2d2-977950b95d61-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.154178 4735 generic.go:334] "Generic (PLEG): container finished" podID="eba24811-9f0a-486b-b2d2-977950b95d61" containerID="2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885" exitCode=0 Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.154226 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrp27" event={"ID":"eba24811-9f0a-486b-b2d2-977950b95d61","Type":"ContainerDied","Data":"2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885"} Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.154256 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrp27" event={"ID":"eba24811-9f0a-486b-b2d2-977950b95d61","Type":"ContainerDied","Data":"26d6624a7ea31ce1bed1e68058bfa41e0415e688dc0f91fe1914b1d4cb0fc471"} Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.154277 4735 scope.go:117] "RemoveContainer" containerID="2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.154310 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrp27" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.175125 4735 scope.go:117] "RemoveContainer" containerID="f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.178351 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rrp27"] Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.186770 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rrp27"] Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.204706 4735 scope.go:117] "RemoveContainer" containerID="6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.219368 4735 scope.go:117] "RemoveContainer" containerID="2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885" Feb 15 20:29:47 crc kubenswrapper[4735]: E0215 20:29:47.219724 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885\": container with ID starting with 2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885 not found: ID does not exist" containerID="2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.219781 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885"} err="failed to get container status \"2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885\": rpc error: code = NotFound desc = could not find container \"2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885\": container with ID starting with 2e5e164b42eaeaf2d72be6265ced89443ea28364efb4df73ec4b12e13f6cb885 not found: ID does not exist" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.219808 4735 scope.go:117] "RemoveContainer" containerID="f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988" Feb 15 20:29:47 crc kubenswrapper[4735]: E0215 20:29:47.220544 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988\": container with ID starting with f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988 not found: ID does not exist" containerID="f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.220594 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988"} err="failed to get container status \"f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988\": rpc error: code = NotFound desc = could not find container \"f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988\": container with ID starting with f847871ee138b8e2140bf1b14f43064bfa5f0f539116b372d9f9e700c096f988 not found: ID does not exist" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.220627 4735 scope.go:117] "RemoveContainer" containerID="6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a" Feb 15 20:29:47 crc kubenswrapper[4735]: E0215 20:29:47.221203 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a\": container with ID starting with 6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a not found: ID does not exist" containerID="6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a" Feb 15 20:29:47 crc kubenswrapper[4735]: I0215 20:29:47.221238 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a"} err="failed to get container status \"6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a\": rpc error: code = NotFound desc = could not find container \"6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a\": container with ID starting with 6ae40d926183f4a45f8b1873bf247c49394612625c89e681ea9b327ae39c269a not found: ID does not exist" Feb 15 20:29:48 crc kubenswrapper[4735]: I0215 20:29:48.893223 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" path="/var/lib/kubelet/pods/eba24811-9f0a-486b-b2d2-977950b95d61/volumes" Feb 15 20:29:49 crc kubenswrapper[4735]: I0215 20:29:49.679413 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:29:49 crc kubenswrapper[4735]: I0215 20:29:49.679468 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:29:49 crc kubenswrapper[4735]: I0215 20:29:49.679506 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:29:49 crc kubenswrapper[4735]: I0215 20:29:49.680061 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8137201feee74b50c12e4ac603bd7a608ffb1bf6f8ae2e8df072ae67e560dfe2"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:29:49 crc kubenswrapper[4735]: I0215 20:29:49.680116 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://8137201feee74b50c12e4ac603bd7a608ffb1bf6f8ae2e8df072ae67e560dfe2" gracePeriod=600 Feb 15 20:29:50 crc kubenswrapper[4735]: I0215 20:29:50.173112 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="8137201feee74b50c12e4ac603bd7a608ffb1bf6f8ae2e8df072ae67e560dfe2" exitCode=0 Feb 15 20:29:50 crc kubenswrapper[4735]: I0215 20:29:50.173191 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"8137201feee74b50c12e4ac603bd7a608ffb1bf6f8ae2e8df072ae67e560dfe2"} Feb 15 20:29:50 crc kubenswrapper[4735]: I0215 20:29:50.173447 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"776a88a7dd582588d5537251275510c6e25c50e714d7281e0f68c913b0f4fe42"} Feb 15 20:29:50 crc kubenswrapper[4735]: I0215 20:29:50.173469 4735 scope.go:117] "RemoveContainer" containerID="925e7518253cb6621d6125aa829f6d9222f7f5dddb79f1eba385c5daf1d88f7d" Feb 15 20:29:57 crc kubenswrapper[4735]: I0215 20:29:57.489449 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7f47cf9495-7nwhm" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.158636 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr"] Feb 15 20:30:00 crc kubenswrapper[4735]: E0215 20:30:00.159061 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" containerName="registry-server" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.159073 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" containerName="registry-server" Feb 15 20:30:00 crc kubenswrapper[4735]: E0215 20:30:00.159082 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" containerName="extract-content" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.159088 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" containerName="extract-content" Feb 15 20:30:00 crc kubenswrapper[4735]: E0215 20:30:00.159097 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" containerName="extract-utilities" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.159103 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" containerName="extract-utilities" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.159214 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="eba24811-9f0a-486b-b2d2-977950b95d61" containerName="registry-server" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.159610 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.161768 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.162003 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.211090 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr"] Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.280741 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-secret-volume\") pod \"collect-profiles-29519790-dkjpr\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.280795 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-config-volume\") pod \"collect-profiles-29519790-dkjpr\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.280826 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g27c4\" (UniqueName: \"kubernetes.io/projected/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-kube-api-access-g27c4\") pod \"collect-profiles-29519790-dkjpr\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.382675 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-secret-volume\") pod \"collect-profiles-29519790-dkjpr\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.382918 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-config-volume\") pod \"collect-profiles-29519790-dkjpr\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.383049 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g27c4\" (UniqueName: \"kubernetes.io/projected/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-kube-api-access-g27c4\") pod \"collect-profiles-29519790-dkjpr\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.383933 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-config-volume\") pod \"collect-profiles-29519790-dkjpr\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.396788 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-secret-volume\") pod \"collect-profiles-29519790-dkjpr\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.407562 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g27c4\" (UniqueName: \"kubernetes.io/projected/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-kube-api-access-g27c4\") pod \"collect-profiles-29519790-dkjpr\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.482182 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:00 crc kubenswrapper[4735]: I0215 20:30:00.779875 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr"] Feb 15 20:30:00 crc kubenswrapper[4735]: W0215 20:30:00.789124 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab473fa2_1b7e_4ea8_ac98_8d883266e2ca.slice/crio-9e8cf107dca13173063627cae66f3b436e012af7c96b5691e7f1d7469b427329 WatchSource:0}: Error finding container 9e8cf107dca13173063627cae66f3b436e012af7c96b5691e7f1d7469b427329: Status 404 returned error can't find the container with id 9e8cf107dca13173063627cae66f3b436e012af7c96b5691e7f1d7469b427329 Feb 15 20:30:01 crc kubenswrapper[4735]: I0215 20:30:01.275683 4735 generic.go:334] "Generic (PLEG): container finished" podID="ab473fa2-1b7e-4ea8-ac98-8d883266e2ca" containerID="fa0e6b2b0b04222955ece57389214bb43f2762b8ef5b9fd83c15651879bf9ede" exitCode=0 Feb 15 20:30:01 crc kubenswrapper[4735]: I0215 20:30:01.275724 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" event={"ID":"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca","Type":"ContainerDied","Data":"fa0e6b2b0b04222955ece57389214bb43f2762b8ef5b9fd83c15651879bf9ede"} Feb 15 20:30:01 crc kubenswrapper[4735]: I0215 20:30:01.275747 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" event={"ID":"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca","Type":"ContainerStarted","Data":"9e8cf107dca13173063627cae66f3b436e012af7c96b5691e7f1d7469b427329"} Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.521202 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.610936 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g27c4\" (UniqueName: \"kubernetes.io/projected/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-kube-api-access-g27c4\") pod \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.611067 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-config-volume\") pod \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.611129 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-secret-volume\") pod \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\" (UID: \"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca\") " Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.613902 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-config-volume" (OuterVolumeSpecName: "config-volume") pod "ab473fa2-1b7e-4ea8-ac98-8d883266e2ca" (UID: "ab473fa2-1b7e-4ea8-ac98-8d883266e2ca"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.616187 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ab473fa2-1b7e-4ea8-ac98-8d883266e2ca" (UID: "ab473fa2-1b7e-4ea8-ac98-8d883266e2ca"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.618021 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-kube-api-access-g27c4" (OuterVolumeSpecName: "kube-api-access-g27c4") pod "ab473fa2-1b7e-4ea8-ac98-8d883266e2ca" (UID: "ab473fa2-1b7e-4ea8-ac98-8d883266e2ca"). InnerVolumeSpecName "kube-api-access-g27c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.712397 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.712433 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g27c4\" (UniqueName: \"kubernetes.io/projected/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-kube-api-access-g27c4\") on node \"crc\" DevicePath \"\"" Feb 15 20:30:02 crc kubenswrapper[4735]: I0215 20:30:02.712444 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca-config-volume\") on node \"crc\" DevicePath \"\"" Feb 15 20:30:03 crc kubenswrapper[4735]: I0215 20:30:03.289253 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" event={"ID":"ab473fa2-1b7e-4ea8-ac98-8d883266e2ca","Type":"ContainerDied","Data":"9e8cf107dca13173063627cae66f3b436e012af7c96b5691e7f1d7469b427329"} Feb 15 20:30:03 crc kubenswrapper[4735]: I0215 20:30:03.289571 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e8cf107dca13173063627cae66f3b436e012af7c96b5691e7f1d7469b427329" Feb 15 20:30:03 crc kubenswrapper[4735]: I0215 20:30:03.289286 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.115242 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-79df65bf49-hzgd7" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.832556 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-g5jlv"] Feb 15 20:30:17 crc kubenswrapper[4735]: E0215 20:30:17.832760 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab473fa2-1b7e-4ea8-ac98-8d883266e2ca" containerName="collect-profiles" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.832775 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab473fa2-1b7e-4ea8-ac98-8d883266e2ca" containerName="collect-profiles" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.832913 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab473fa2-1b7e-4ea8-ac98-8d883266e2ca" containerName="collect-profiles" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.834628 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.840189 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-xxkbt" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.840499 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.840657 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.848593 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29"] Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.849738 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.852900 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.869182 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29"] Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.951381 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-9tpjm"] Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.953066 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9tpjm" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.965704 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.965913 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.966262 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-z4h69" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.966415 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.970244 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-7xh6x"] Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.971053 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.974998 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 15 20:30:17 crc kubenswrapper[4735]: I0215 20:30:17.981486 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-7xh6x"] Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.008873 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-reloader\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.008984 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99caf845-7f1f-4a29-a8e5-c939ffb15c9e-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7km29\" (UID: \"99caf845-7f1f-4a29-a8e5-c939ffb15c9e\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.009018 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-metrics\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.009070 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skzt7\" (UniqueName: \"kubernetes.io/projected/d5d3c0c9-8711-4346-9591-24807ee906ef-kube-api-access-skzt7\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.009108 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5d3c0c9-8711-4346-9591-24807ee906ef-metrics-certs\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.009146 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d5d3c0c9-8711-4346-9591-24807ee906ef-frr-startup\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.009162 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-frr-sockets\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.009179 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-frr-conf\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.009206 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl4qq\" (UniqueName: \"kubernetes.io/projected/99caf845-7f1f-4a29-a8e5-c939ffb15c9e-kube-api-access-tl4qq\") pod \"frr-k8s-webhook-server-78b44bf5bb-7km29\" (UID: \"99caf845-7f1f-4a29-a8e5-c939ffb15c9e\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110557 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5d3c0c9-8711-4346-9591-24807ee906ef-metrics-certs\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110615 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d5d3c0c9-8711-4346-9591-24807ee906ef-frr-startup\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110637 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-frr-sockets\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110651 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-frr-conf\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110681 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl4qq\" (UniqueName: \"kubernetes.io/projected/99caf845-7f1f-4a29-a8e5-c939ffb15c9e-kube-api-access-tl4qq\") pod \"frr-k8s-webhook-server-78b44bf5bb-7km29\" (UID: \"99caf845-7f1f-4a29-a8e5-c939ffb15c9e\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110710 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wth6\" (UniqueName: \"kubernetes.io/projected/ca5b18f3-83a4-497d-913b-e0539a6991b5-kube-api-access-6wth6\") pod \"controller-69bbfbf88f-7xh6x\" (UID: \"ca5b18f3-83a4-497d-913b-e0539a6991b5\") " pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110731 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-reloader\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110748 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99caf845-7f1f-4a29-a8e5-c939ffb15c9e-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7km29\" (UID: \"99caf845-7f1f-4a29-a8e5-c939ffb15c9e\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110775 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ec4ea5be-56b7-42e2-b462-a880d12145af-metallb-excludel2\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110797 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec4ea5be-56b7-42e2-b462-a880d12145af-memberlist\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110813 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-metrics\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110831 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca5b18f3-83a4-497d-913b-e0539a6991b5-cert\") pod \"controller-69bbfbf88f-7xh6x\" (UID: \"ca5b18f3-83a4-497d-913b-e0539a6991b5\") " pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110849 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec4ea5be-56b7-42e2-b462-a880d12145af-metrics-certs\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110868 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46czx\" (UniqueName: \"kubernetes.io/projected/ec4ea5be-56b7-42e2-b462-a880d12145af-kube-api-access-46czx\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110889 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skzt7\" (UniqueName: \"kubernetes.io/projected/d5d3c0c9-8711-4346-9591-24807ee906ef-kube-api-access-skzt7\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.110915 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca5b18f3-83a4-497d-913b-e0539a6991b5-metrics-certs\") pod \"controller-69bbfbf88f-7xh6x\" (UID: \"ca5b18f3-83a4-497d-913b-e0539a6991b5\") " pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.111319 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-frr-sockets\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.111492 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-frr-conf\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.111690 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-metrics\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.111688 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d5d3c0c9-8711-4346-9591-24807ee906ef-frr-startup\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: E0215 20:30:18.111903 4735 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Feb 15 20:30:18 crc kubenswrapper[4735]: E0215 20:30:18.112084 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/99caf845-7f1f-4a29-a8e5-c939ffb15c9e-cert podName:99caf845-7f1f-4a29-a8e5-c939ffb15c9e nodeName:}" failed. No retries permitted until 2026-02-15 20:30:18.612057606 +0000 UTC m=+826.478073309 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/99caf845-7f1f-4a29-a8e5-c939ffb15c9e-cert") pod "frr-k8s-webhook-server-78b44bf5bb-7km29" (UID: "99caf845-7f1f-4a29-a8e5-c939ffb15c9e") : secret "frr-k8s-webhook-server-cert" not found Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.111911 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d5d3c0c9-8711-4346-9591-24807ee906ef-reloader\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.117357 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5d3c0c9-8711-4346-9591-24807ee906ef-metrics-certs\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.137322 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skzt7\" (UniqueName: \"kubernetes.io/projected/d5d3c0c9-8711-4346-9591-24807ee906ef-kube-api-access-skzt7\") pod \"frr-k8s-g5jlv\" (UID: \"d5d3c0c9-8711-4346-9591-24807ee906ef\") " pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.138604 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl4qq\" (UniqueName: \"kubernetes.io/projected/99caf845-7f1f-4a29-a8e5-c939ffb15c9e-kube-api-access-tl4qq\") pod \"frr-k8s-webhook-server-78b44bf5bb-7km29\" (UID: \"99caf845-7f1f-4a29-a8e5-c939ffb15c9e\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.168175 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.212508 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ec4ea5be-56b7-42e2-b462-a880d12145af-metallb-excludel2\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.212548 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec4ea5be-56b7-42e2-b462-a880d12145af-memberlist\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.212584 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca5b18f3-83a4-497d-913b-e0539a6991b5-cert\") pod \"controller-69bbfbf88f-7xh6x\" (UID: \"ca5b18f3-83a4-497d-913b-e0539a6991b5\") " pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.212602 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec4ea5be-56b7-42e2-b462-a880d12145af-metrics-certs\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.212621 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46czx\" (UniqueName: \"kubernetes.io/projected/ec4ea5be-56b7-42e2-b462-a880d12145af-kube-api-access-46czx\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.212665 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca5b18f3-83a4-497d-913b-e0539a6991b5-metrics-certs\") pod \"controller-69bbfbf88f-7xh6x\" (UID: \"ca5b18f3-83a4-497d-913b-e0539a6991b5\") " pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.212707 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wth6\" (UniqueName: \"kubernetes.io/projected/ca5b18f3-83a4-497d-913b-e0539a6991b5-kube-api-access-6wth6\") pod \"controller-69bbfbf88f-7xh6x\" (UID: \"ca5b18f3-83a4-497d-913b-e0539a6991b5\") " pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.214268 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ec4ea5be-56b7-42e2-b462-a880d12145af-metallb-excludel2\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: E0215 20:30:18.214370 4735 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 15 20:30:18 crc kubenswrapper[4735]: E0215 20:30:18.214408 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec4ea5be-56b7-42e2-b462-a880d12145af-memberlist podName:ec4ea5be-56b7-42e2-b462-a880d12145af nodeName:}" failed. No retries permitted until 2026-02-15 20:30:18.71439667 +0000 UTC m=+826.580412293 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ec4ea5be-56b7-42e2-b462-a880d12145af-memberlist") pod "speaker-9tpjm" (UID: "ec4ea5be-56b7-42e2-b462-a880d12145af") : secret "metallb-memberlist" not found Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.216076 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca5b18f3-83a4-497d-913b-e0539a6991b5-cert\") pod \"controller-69bbfbf88f-7xh6x\" (UID: \"ca5b18f3-83a4-497d-913b-e0539a6991b5\") " pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.218024 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca5b18f3-83a4-497d-913b-e0539a6991b5-metrics-certs\") pod \"controller-69bbfbf88f-7xh6x\" (UID: \"ca5b18f3-83a4-497d-913b-e0539a6991b5\") " pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.222428 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec4ea5be-56b7-42e2-b462-a880d12145af-metrics-certs\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.243205 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46czx\" (UniqueName: \"kubernetes.io/projected/ec4ea5be-56b7-42e2-b462-a880d12145af-kube-api-access-46czx\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.243432 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wth6\" (UniqueName: \"kubernetes.io/projected/ca5b18f3-83a4-497d-913b-e0539a6991b5-kube-api-access-6wth6\") pod \"controller-69bbfbf88f-7xh6x\" (UID: \"ca5b18f3-83a4-497d-913b-e0539a6991b5\") " pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.292197 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.380906 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerStarted","Data":"fdc21d8d8085e91d8f075e3531e5ccf9811a276d91f6268deaf3981c9ab4fa70"} Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.553823 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-7xh6x"] Feb 15 20:30:18 crc kubenswrapper[4735]: W0215 20:30:18.556662 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca5b18f3_83a4_497d_913b_e0539a6991b5.slice/crio-a8aa51cc51039ac48b2a5fabae6f96a07311157c6209e3a050c1724a27b20779 WatchSource:0}: Error finding container a8aa51cc51039ac48b2a5fabae6f96a07311157c6209e3a050c1724a27b20779: Status 404 returned error can't find the container with id a8aa51cc51039ac48b2a5fabae6f96a07311157c6209e3a050c1724a27b20779 Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.632517 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99caf845-7f1f-4a29-a8e5-c939ffb15c9e-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7km29\" (UID: \"99caf845-7f1f-4a29-a8e5-c939ffb15c9e\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.636886 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99caf845-7f1f-4a29-a8e5-c939ffb15c9e-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7km29\" (UID: \"99caf845-7f1f-4a29-a8e5-c939ffb15c9e\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.733429 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec4ea5be-56b7-42e2-b462-a880d12145af-memberlist\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.738183 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec4ea5be-56b7-42e2-b462-a880d12145af-memberlist\") pod \"speaker-9tpjm\" (UID: \"ec4ea5be-56b7-42e2-b462-a880d12145af\") " pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.774673 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.880561 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9tpjm" Feb 15 20:30:18 crc kubenswrapper[4735]: W0215 20:30:18.918883 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec4ea5be_56b7_42e2_b462_a880d12145af.slice/crio-eaf856f022d12156398452e4023d6f097bfa5364a3ce6088115a2aee472f0e0d WatchSource:0}: Error finding container eaf856f022d12156398452e4023d6f097bfa5364a3ce6088115a2aee472f0e0d: Status 404 returned error can't find the container with id eaf856f022d12156398452e4023d6f097bfa5364a3ce6088115a2aee472f0e0d Feb 15 20:30:18 crc kubenswrapper[4735]: I0215 20:30:18.980164 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29"] Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.387715 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-7xh6x" event={"ID":"ca5b18f3-83a4-497d-913b-e0539a6991b5","Type":"ContainerStarted","Data":"34b58a47e2ee88ac2f1d7a84b7fe0fc2615398cbcfa71ec97f378184a68c779d"} Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.388060 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-7xh6x" event={"ID":"ca5b18f3-83a4-497d-913b-e0539a6991b5","Type":"ContainerStarted","Data":"10451046a7eb9f90c75dd097718b51139b6798f3c4aaa6acfe7813c33f71e637"} Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.388083 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.388095 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-7xh6x" event={"ID":"ca5b18f3-83a4-497d-913b-e0539a6991b5","Type":"ContainerStarted","Data":"a8aa51cc51039ac48b2a5fabae6f96a07311157c6209e3a050c1724a27b20779"} Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.390616 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9tpjm" event={"ID":"ec4ea5be-56b7-42e2-b462-a880d12145af","Type":"ContainerStarted","Data":"573bdadb7baf53ba100e0c6288c29ccc9bf250095236207de045b7c82b616696"} Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.390655 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9tpjm" event={"ID":"ec4ea5be-56b7-42e2-b462-a880d12145af","Type":"ContainerStarted","Data":"5185c1cf8ddb21820d8fd0c159bffd41e7ab8d1669b1485e5d26ed410f3d0ab8"} Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.390667 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9tpjm" event={"ID":"ec4ea5be-56b7-42e2-b462-a880d12145af","Type":"ContainerStarted","Data":"eaf856f022d12156398452e4023d6f097bfa5364a3ce6088115a2aee472f0e0d"} Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.391287 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-9tpjm" Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.392054 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" event={"ID":"99caf845-7f1f-4a29-a8e5-c939ffb15c9e","Type":"ContainerStarted","Data":"75b2bfa068e5dd4987118f1ec7ebe99e457d19b16176e9e18f52855f0f5647bc"} Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.406463 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-7xh6x" podStartSLOduration=2.406447894 podStartE2EDuration="2.406447894s" podCreationTimestamp="2026-02-15 20:30:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:30:19.402087571 +0000 UTC m=+827.268103194" watchObservedRunningTime="2026-02-15 20:30:19.406447894 +0000 UTC m=+827.272463517" Feb 15 20:30:19 crc kubenswrapper[4735]: I0215 20:30:19.428297 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-9tpjm" podStartSLOduration=2.428278925 podStartE2EDuration="2.428278925s" podCreationTimestamp="2026-02-15 20:30:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:30:19.426122348 +0000 UTC m=+827.292137971" watchObservedRunningTime="2026-02-15 20:30:19.428278925 +0000 UTC m=+827.294294548" Feb 15 20:30:26 crc kubenswrapper[4735]: I0215 20:30:26.462250 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" event={"ID":"99caf845-7f1f-4a29-a8e5-c939ffb15c9e","Type":"ContainerStarted","Data":"cfda4edd2b72b177bfa771c6486773d392d50c359167a162ff93c6157da86e98"} Feb 15 20:30:26 crc kubenswrapper[4735]: I0215 20:30:26.462764 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:26 crc kubenswrapper[4735]: I0215 20:30:26.464736 4735 generic.go:334] "Generic (PLEG): container finished" podID="d5d3c0c9-8711-4346-9591-24807ee906ef" containerID="740d8d2f4e7cd6d16286799ab50fffe725fbbece6239316b0af4540b37f79a01" exitCode=0 Feb 15 20:30:26 crc kubenswrapper[4735]: I0215 20:30:26.464786 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerDied","Data":"740d8d2f4e7cd6d16286799ab50fffe725fbbece6239316b0af4540b37f79a01"} Feb 15 20:30:26 crc kubenswrapper[4735]: I0215 20:30:26.484158 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" podStartSLOduration=2.327417089 podStartE2EDuration="9.484142201s" podCreationTimestamp="2026-02-15 20:30:17 +0000 UTC" firstStartedPulling="2026-02-15 20:30:18.992368915 +0000 UTC m=+826.858384538" lastFinishedPulling="2026-02-15 20:30:26.149094027 +0000 UTC m=+834.015109650" observedRunningTime="2026-02-15 20:30:26.479679364 +0000 UTC m=+834.345694987" watchObservedRunningTime="2026-02-15 20:30:26.484142201 +0000 UTC m=+834.350157824" Feb 15 20:30:27 crc kubenswrapper[4735]: I0215 20:30:27.473392 4735 generic.go:334] "Generic (PLEG): container finished" podID="d5d3c0c9-8711-4346-9591-24807ee906ef" containerID="85e1a7198494e65efa9c4f986d2b95e4d51b89f895704ed99bb71a8f52b7727e" exitCode=0 Feb 15 20:30:27 crc kubenswrapper[4735]: I0215 20:30:27.473499 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerDied","Data":"85e1a7198494e65efa9c4f986d2b95e4d51b89f895704ed99bb71a8f52b7727e"} Feb 15 20:30:28 crc kubenswrapper[4735]: I0215 20:30:28.298269 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-7xh6x" Feb 15 20:30:28 crc kubenswrapper[4735]: I0215 20:30:28.481019 4735 generic.go:334] "Generic (PLEG): container finished" podID="d5d3c0c9-8711-4346-9591-24807ee906ef" containerID="861d2911ef3bbb3176560f0eb86968c47304c88d8f0bf1be5173ff5c48f540f1" exitCode=0 Feb 15 20:30:28 crc kubenswrapper[4735]: I0215 20:30:28.481061 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerDied","Data":"861d2911ef3bbb3176560f0eb86968c47304c88d8f0bf1be5173ff5c48f540f1"} Feb 15 20:30:29 crc kubenswrapper[4735]: I0215 20:30:29.492605 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerStarted","Data":"91584123b06cdab7cc16d13d634bce0aebe1e7ba0413963069797fdd1e2abbc4"} Feb 15 20:30:29 crc kubenswrapper[4735]: I0215 20:30:29.492988 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerStarted","Data":"be47472ec96469ff68740f62f8b71a21cff797a2f360309b5aaa2e89ac2a45cb"} Feb 15 20:30:29 crc kubenswrapper[4735]: I0215 20:30:29.493232 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:29 crc kubenswrapper[4735]: I0215 20:30:29.493266 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerStarted","Data":"a2d7e60f5a513f1e358c8acbe1cac51cd917b09154ada537318c3753f640f863"} Feb 15 20:30:29 crc kubenswrapper[4735]: I0215 20:30:29.493287 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerStarted","Data":"9440721c673d1ac97150f6f4a504021d0128ff94925f8457a069023d5d7f8c2b"} Feb 15 20:30:29 crc kubenswrapper[4735]: I0215 20:30:29.493329 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerStarted","Data":"3a969cf5633a6ee9ebf32097460d8a524cdfae7d424ae1c87e9ba5de6fde354d"} Feb 15 20:30:29 crc kubenswrapper[4735]: I0215 20:30:29.493340 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g5jlv" event={"ID":"d5d3c0c9-8711-4346-9591-24807ee906ef","Type":"ContainerStarted","Data":"8d6302154277a52e494a656169c769d280eb7439d655d9769cdef5ba5f945477"} Feb 15 20:30:29 crc kubenswrapper[4735]: I0215 20:30:29.529379 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-g5jlv" podStartSLOduration=4.657635051 podStartE2EDuration="12.529348193s" podCreationTimestamp="2026-02-15 20:30:17 +0000 UTC" firstStartedPulling="2026-02-15 20:30:18.30819835 +0000 UTC m=+826.174213973" lastFinishedPulling="2026-02-15 20:30:26.179911482 +0000 UTC m=+834.045927115" observedRunningTime="2026-02-15 20:30:29.526384555 +0000 UTC m=+837.392400178" watchObservedRunningTime="2026-02-15 20:30:29.529348193 +0000 UTC m=+837.395363856" Feb 15 20:30:33 crc kubenswrapper[4735]: I0215 20:30:33.169128 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:33 crc kubenswrapper[4735]: I0215 20:30:33.235010 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:38 crc kubenswrapper[4735]: I0215 20:30:38.173138 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-g5jlv" Feb 15 20:30:38 crc kubenswrapper[4735]: I0215 20:30:38.780506 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7km29" Feb 15 20:30:38 crc kubenswrapper[4735]: I0215 20:30:38.884396 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-9tpjm" Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.572991 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-gzslz"] Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.573665 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gzslz" Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.582241 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.584217 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.585253 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-td9bx" Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.653986 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-gzslz"] Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.680886 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4czhx\" (UniqueName: \"kubernetes.io/projected/f74cbdd5-7979-4d6a-8c04-f6377965c5cf-kube-api-access-4czhx\") pod \"openstack-operator-index-gzslz\" (UID: \"f74cbdd5-7979-4d6a-8c04-f6377965c5cf\") " pod="openstack-operators/openstack-operator-index-gzslz" Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.782234 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4czhx\" (UniqueName: \"kubernetes.io/projected/f74cbdd5-7979-4d6a-8c04-f6377965c5cf-kube-api-access-4czhx\") pod \"openstack-operator-index-gzslz\" (UID: \"f74cbdd5-7979-4d6a-8c04-f6377965c5cf\") " pod="openstack-operators/openstack-operator-index-gzslz" Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.799038 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4czhx\" (UniqueName: \"kubernetes.io/projected/f74cbdd5-7979-4d6a-8c04-f6377965c5cf-kube-api-access-4czhx\") pod \"openstack-operator-index-gzslz\" (UID: \"f74cbdd5-7979-4d6a-8c04-f6377965c5cf\") " pod="openstack-operators/openstack-operator-index-gzslz" Feb 15 20:30:41 crc kubenswrapper[4735]: I0215 20:30:41.898863 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gzslz" Feb 15 20:30:42 crc kubenswrapper[4735]: I0215 20:30:42.376276 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-gzslz"] Feb 15 20:30:42 crc kubenswrapper[4735]: W0215 20:30:42.383823 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf74cbdd5_7979_4d6a_8c04_f6377965c5cf.slice/crio-425cd8ea25c74f91327fa5a950dbbaefecb9b317d7cc0c50723c7ceebcf51cb5 WatchSource:0}: Error finding container 425cd8ea25c74f91327fa5a950dbbaefecb9b317d7cc0c50723c7ceebcf51cb5: Status 404 returned error can't find the container with id 425cd8ea25c74f91327fa5a950dbbaefecb9b317d7cc0c50723c7ceebcf51cb5 Feb 15 20:30:42 crc kubenswrapper[4735]: I0215 20:30:42.583432 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gzslz" event={"ID":"f74cbdd5-7979-4d6a-8c04-f6377965c5cf","Type":"ContainerStarted","Data":"425cd8ea25c74f91327fa5a950dbbaefecb9b317d7cc0c50723c7ceebcf51cb5"} Feb 15 20:30:44 crc kubenswrapper[4735]: I0215 20:30:44.951318 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-gzslz"] Feb 15 20:30:45 crc kubenswrapper[4735]: I0215 20:30:45.562560 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-ldbt6"] Feb 15 20:30:45 crc kubenswrapper[4735]: I0215 20:30:45.564112 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ldbt6" Feb 15 20:30:45 crc kubenswrapper[4735]: I0215 20:30:45.580407 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ldbt6"] Feb 15 20:30:45 crc kubenswrapper[4735]: I0215 20:30:45.609197 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gzslz" event={"ID":"f74cbdd5-7979-4d6a-8c04-f6377965c5cf","Type":"ContainerStarted","Data":"aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc"} Feb 15 20:30:45 crc kubenswrapper[4735]: I0215 20:30:45.645596 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-gzslz" podStartSLOduration=2.360677043 podStartE2EDuration="4.645540188s" podCreationTimestamp="2026-02-15 20:30:41 +0000 UTC" firstStartedPulling="2026-02-15 20:30:42.387592169 +0000 UTC m=+850.253607792" lastFinishedPulling="2026-02-15 20:30:44.672455314 +0000 UTC m=+852.538470937" observedRunningTime="2026-02-15 20:30:45.634662674 +0000 UTC m=+853.500678307" watchObservedRunningTime="2026-02-15 20:30:45.645540188 +0000 UTC m=+853.511555831" Feb 15 20:30:45 crc kubenswrapper[4735]: I0215 20:30:45.736593 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6brlf\" (UniqueName: \"kubernetes.io/projected/32c1d680-92cf-4c49-93b9-17bb17aaf623-kube-api-access-6brlf\") pod \"openstack-operator-index-ldbt6\" (UID: \"32c1d680-92cf-4c49-93b9-17bb17aaf623\") " pod="openstack-operators/openstack-operator-index-ldbt6" Feb 15 20:30:45 crc kubenswrapper[4735]: I0215 20:30:45.838279 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6brlf\" (UniqueName: \"kubernetes.io/projected/32c1d680-92cf-4c49-93b9-17bb17aaf623-kube-api-access-6brlf\") pod \"openstack-operator-index-ldbt6\" (UID: \"32c1d680-92cf-4c49-93b9-17bb17aaf623\") " pod="openstack-operators/openstack-operator-index-ldbt6" Feb 15 20:30:45 crc kubenswrapper[4735]: I0215 20:30:45.858787 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6brlf\" (UniqueName: \"kubernetes.io/projected/32c1d680-92cf-4c49-93b9-17bb17aaf623-kube-api-access-6brlf\") pod \"openstack-operator-index-ldbt6\" (UID: \"32c1d680-92cf-4c49-93b9-17bb17aaf623\") " pod="openstack-operators/openstack-operator-index-ldbt6" Feb 15 20:30:45 crc kubenswrapper[4735]: I0215 20:30:45.889223 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ldbt6" Feb 15 20:30:46 crc kubenswrapper[4735]: I0215 20:30:46.375350 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ldbt6"] Feb 15 20:30:46 crc kubenswrapper[4735]: W0215 20:30:46.380803 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32c1d680_92cf_4c49_93b9_17bb17aaf623.slice/crio-a1b3e7794b8f8de0c482234fa5f088ad3d325c79f947a8e3494ef83e43bb3fb4 WatchSource:0}: Error finding container a1b3e7794b8f8de0c482234fa5f088ad3d325c79f947a8e3494ef83e43bb3fb4: Status 404 returned error can't find the container with id a1b3e7794b8f8de0c482234fa5f088ad3d325c79f947a8e3494ef83e43bb3fb4 Feb 15 20:30:46 crc kubenswrapper[4735]: I0215 20:30:46.618754 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-gzslz" podUID="f74cbdd5-7979-4d6a-8c04-f6377965c5cf" containerName="registry-server" containerID="cri-o://aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc" gracePeriod=2 Feb 15 20:30:46 crc kubenswrapper[4735]: I0215 20:30:46.619844 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ldbt6" event={"ID":"32c1d680-92cf-4c49-93b9-17bb17aaf623","Type":"ContainerStarted","Data":"048fe5b5d4d11d03dfea67a4f1af4f45dec20118dca992e6b37da9b1a8f32614"} Feb 15 20:30:46 crc kubenswrapper[4735]: I0215 20:30:46.619940 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ldbt6" event={"ID":"32c1d680-92cf-4c49-93b9-17bb17aaf623","Type":"ContainerStarted","Data":"a1b3e7794b8f8de0c482234fa5f088ad3d325c79f947a8e3494ef83e43bb3fb4"} Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.023706 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gzslz" Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.043091 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-ldbt6" podStartSLOduration=1.995719125 podStartE2EDuration="2.043073621s" podCreationTimestamp="2026-02-15 20:30:45 +0000 UTC" firstStartedPulling="2026-02-15 20:30:46.387510034 +0000 UTC m=+854.253525677" lastFinishedPulling="2026-02-15 20:30:46.43486454 +0000 UTC m=+854.300880173" observedRunningTime="2026-02-15 20:30:46.639647011 +0000 UTC m=+854.505662674" watchObservedRunningTime="2026-02-15 20:30:47.043073621 +0000 UTC m=+854.909089234" Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.154933 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4czhx\" (UniqueName: \"kubernetes.io/projected/f74cbdd5-7979-4d6a-8c04-f6377965c5cf-kube-api-access-4czhx\") pod \"f74cbdd5-7979-4d6a-8c04-f6377965c5cf\" (UID: \"f74cbdd5-7979-4d6a-8c04-f6377965c5cf\") " Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.169176 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f74cbdd5-7979-4d6a-8c04-f6377965c5cf-kube-api-access-4czhx" (OuterVolumeSpecName: "kube-api-access-4czhx") pod "f74cbdd5-7979-4d6a-8c04-f6377965c5cf" (UID: "f74cbdd5-7979-4d6a-8c04-f6377965c5cf"). InnerVolumeSpecName "kube-api-access-4czhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.256267 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4czhx\" (UniqueName: \"kubernetes.io/projected/f74cbdd5-7979-4d6a-8c04-f6377965c5cf-kube-api-access-4czhx\") on node \"crc\" DevicePath \"\"" Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.626472 4735 generic.go:334] "Generic (PLEG): container finished" podID="f74cbdd5-7979-4d6a-8c04-f6377965c5cf" containerID="aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc" exitCode=0 Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.626515 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gzslz" event={"ID":"f74cbdd5-7979-4d6a-8c04-f6377965c5cf","Type":"ContainerDied","Data":"aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc"} Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.626565 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gzslz" event={"ID":"f74cbdd5-7979-4d6a-8c04-f6377965c5cf","Type":"ContainerDied","Data":"425cd8ea25c74f91327fa5a950dbbaefecb9b317d7cc0c50723c7ceebcf51cb5"} Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.626586 4735 scope.go:117] "RemoveContainer" containerID="aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc" Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.626534 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gzslz" Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.652722 4735 scope.go:117] "RemoveContainer" containerID="aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc" Feb 15 20:30:47 crc kubenswrapper[4735]: E0215 20:30:47.653175 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc\": container with ID starting with aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc not found: ID does not exist" containerID="aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc" Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.653223 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc"} err="failed to get container status \"aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc\": rpc error: code = NotFound desc = could not find container \"aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc\": container with ID starting with aef88176cfc218656f8b752c2f8079b4ddc0257a0032c3b6ad26b2e93ee828bc not found: ID does not exist" Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.669160 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-gzslz"] Feb 15 20:30:47 crc kubenswrapper[4735]: I0215 20:30:47.674616 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-gzslz"] Feb 15 20:30:48 crc kubenswrapper[4735]: I0215 20:30:48.895369 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f74cbdd5-7979-4d6a-8c04-f6377965c5cf" path="/var/lib/kubelet/pods/f74cbdd5-7979-4d6a-8c04-f6377965c5cf/volumes" Feb 15 20:30:55 crc kubenswrapper[4735]: I0215 20:30:55.898162 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-ldbt6" Feb 15 20:30:55 crc kubenswrapper[4735]: I0215 20:30:55.898793 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-ldbt6" Feb 15 20:30:55 crc kubenswrapper[4735]: I0215 20:30:55.949851 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-ldbt6" Feb 15 20:30:56 crc kubenswrapper[4735]: I0215 20:30:56.735420 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-ldbt6" Feb 15 20:30:57 crc kubenswrapper[4735]: I0215 20:30:57.804793 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq"] Feb 15 20:30:57 crc kubenswrapper[4735]: E0215 20:30:57.805852 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f74cbdd5-7979-4d6a-8c04-f6377965c5cf" containerName="registry-server" Feb 15 20:30:57 crc kubenswrapper[4735]: I0215 20:30:57.805879 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f74cbdd5-7979-4d6a-8c04-f6377965c5cf" containerName="registry-server" Feb 15 20:30:57 crc kubenswrapper[4735]: I0215 20:30:57.806203 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f74cbdd5-7979-4d6a-8c04-f6377965c5cf" containerName="registry-server" Feb 15 20:30:57 crc kubenswrapper[4735]: I0215 20:30:57.808049 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:57 crc kubenswrapper[4735]: I0215 20:30:57.811299 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lnwr4" Feb 15 20:30:57 crc kubenswrapper[4735]: I0215 20:30:57.831058 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq"] Feb 15 20:30:57 crc kubenswrapper[4735]: I0215 20:30:57.973532 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-bundle\") pod \"8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:57 crc kubenswrapper[4735]: I0215 20:30:57.973828 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-util\") pod \"8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:57 crc kubenswrapper[4735]: I0215 20:30:57.973979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fgff\" (UniqueName: \"kubernetes.io/projected/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-kube-api-access-8fgff\") pod \"8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.076224 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fgff\" (UniqueName: \"kubernetes.io/projected/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-kube-api-access-8fgff\") pod \"8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.076476 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-bundle\") pod \"8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.076540 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-util\") pod \"8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.077260 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-bundle\") pod \"8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.077478 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-util\") pod \"8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.108971 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fgff\" (UniqueName: \"kubernetes.io/projected/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-kube-api-access-8fgff\") pod \"8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.137212 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.358822 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq"] Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.717612 4735 generic.go:334] "Generic (PLEG): container finished" podID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerID="67a27242821d91ef77b7fdabf4078922f2c1c22beb79465883205211ebe2914f" exitCode=0 Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.717702 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" event={"ID":"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af","Type":"ContainerDied","Data":"67a27242821d91ef77b7fdabf4078922f2c1c22beb79465883205211ebe2914f"} Feb 15 20:30:58 crc kubenswrapper[4735]: I0215 20:30:58.717763 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" event={"ID":"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af","Type":"ContainerStarted","Data":"70eaa4c9406364bdec616ae42cfb57e0090d573278b385abe54b455dd61352ee"} Feb 15 20:30:59 crc kubenswrapper[4735]: I0215 20:30:59.727022 4735 generic.go:334] "Generic (PLEG): container finished" podID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerID="0cad48dce671c990cb4b90156959df17cc3b277261fdd9395072d9ce78d72858" exitCode=0 Feb 15 20:30:59 crc kubenswrapper[4735]: I0215 20:30:59.727106 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" event={"ID":"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af","Type":"ContainerDied","Data":"0cad48dce671c990cb4b90156959df17cc3b277261fdd9395072d9ce78d72858"} Feb 15 20:31:00 crc kubenswrapper[4735]: I0215 20:31:00.749672 4735 generic.go:334] "Generic (PLEG): container finished" podID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerID="53b59380b5e43c4f1271de3c3aedb790383b501647abc2c34c8abce9f9f815a8" exitCode=0 Feb 15 20:31:00 crc kubenswrapper[4735]: I0215 20:31:00.750079 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" event={"ID":"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af","Type":"ContainerDied","Data":"53b59380b5e43c4f1271de3c3aedb790383b501647abc2c34c8abce9f9f815a8"} Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.052147 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.250273 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fgff\" (UniqueName: \"kubernetes.io/projected/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-kube-api-access-8fgff\") pod \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.250420 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-util\") pod \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.250446 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-bundle\") pod \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\" (UID: \"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af\") " Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.251380 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-bundle" (OuterVolumeSpecName: "bundle") pod "8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" (UID: "8ceaf5ed-e64d-4d6c-8990-2184dddaf8af"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.258714 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-kube-api-access-8fgff" (OuterVolumeSpecName: "kube-api-access-8fgff") pod "8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" (UID: "8ceaf5ed-e64d-4d6c-8990-2184dddaf8af"). InnerVolumeSpecName "kube-api-access-8fgff". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.288289 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-util" (OuterVolumeSpecName: "util") pod "8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" (UID: "8ceaf5ed-e64d-4d6c-8990-2184dddaf8af"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.352261 4735 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-util\") on node \"crc\" DevicePath \"\"" Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.352294 4735 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.352304 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fgff\" (UniqueName: \"kubernetes.io/projected/8ceaf5ed-e64d-4d6c-8990-2184dddaf8af-kube-api-access-8fgff\") on node \"crc\" DevicePath \"\"" Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.776218 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" event={"ID":"8ceaf5ed-e64d-4d6c-8990-2184dddaf8af","Type":"ContainerDied","Data":"70eaa4c9406364bdec616ae42cfb57e0090d573278b385abe54b455dd61352ee"} Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.776555 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70eaa4c9406364bdec616ae42cfb57e0090d573278b385abe54b455dd61352ee" Feb 15 20:31:02 crc kubenswrapper[4735]: I0215 20:31:02.776318 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq" Feb 15 20:31:09 crc kubenswrapper[4735]: I0215 20:31:09.888249 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp"] Feb 15 20:31:09 crc kubenswrapper[4735]: E0215 20:31:09.889014 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerName="extract" Feb 15 20:31:09 crc kubenswrapper[4735]: I0215 20:31:09.889025 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerName="extract" Feb 15 20:31:09 crc kubenswrapper[4735]: E0215 20:31:09.889039 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerName="pull" Feb 15 20:31:09 crc kubenswrapper[4735]: I0215 20:31:09.889045 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerName="pull" Feb 15 20:31:09 crc kubenswrapper[4735]: E0215 20:31:09.889059 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerName="util" Feb 15 20:31:09 crc kubenswrapper[4735]: I0215 20:31:09.889066 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerName="util" Feb 15 20:31:09 crc kubenswrapper[4735]: I0215 20:31:09.889205 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ceaf5ed-e64d-4d6c-8990-2184dddaf8af" containerName="extract" Feb 15 20:31:09 crc kubenswrapper[4735]: I0215 20:31:09.889589 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" Feb 15 20:31:09 crc kubenswrapper[4735]: I0215 20:31:09.897210 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-xgp5q" Feb 15 20:31:09 crc kubenswrapper[4735]: I0215 20:31:09.947202 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp"] Feb 15 20:31:09 crc kubenswrapper[4735]: I0215 20:31:09.955045 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb8h5\" (UniqueName: \"kubernetes.io/projected/e8c2edc3-d324-4b8e-9831-fcf924f6a655-kube-api-access-lb8h5\") pod \"openstack-operator-controller-init-58498c75bf-2kzsp\" (UID: \"e8c2edc3-d324-4b8e-9831-fcf924f6a655\") " pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" Feb 15 20:31:10 crc kubenswrapper[4735]: I0215 20:31:10.056489 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb8h5\" (UniqueName: \"kubernetes.io/projected/e8c2edc3-d324-4b8e-9831-fcf924f6a655-kube-api-access-lb8h5\") pod \"openstack-operator-controller-init-58498c75bf-2kzsp\" (UID: \"e8c2edc3-d324-4b8e-9831-fcf924f6a655\") " pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" Feb 15 20:31:10 crc kubenswrapper[4735]: I0215 20:31:10.079272 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb8h5\" (UniqueName: \"kubernetes.io/projected/e8c2edc3-d324-4b8e-9831-fcf924f6a655-kube-api-access-lb8h5\") pod \"openstack-operator-controller-init-58498c75bf-2kzsp\" (UID: \"e8c2edc3-d324-4b8e-9831-fcf924f6a655\") " pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" Feb 15 20:31:10 crc kubenswrapper[4735]: I0215 20:31:10.206737 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" Feb 15 20:31:10 crc kubenswrapper[4735]: I0215 20:31:10.451705 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp"] Feb 15 20:31:10 crc kubenswrapper[4735]: I0215 20:31:10.838125 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" event={"ID":"e8c2edc3-d324-4b8e-9831-fcf924f6a655","Type":"ContainerStarted","Data":"db601791179387cdaafacc494d67c50b7369db13ae18a1641fd8561f407f3b90"} Feb 15 20:31:15 crc kubenswrapper[4735]: I0215 20:31:15.870673 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" event={"ID":"e8c2edc3-d324-4b8e-9831-fcf924f6a655","Type":"ContainerStarted","Data":"36395148c1cabbcc86c611f62df9da18b6b3ff0fe77b879b767f5f52d5f8ea8f"} Feb 15 20:31:15 crc kubenswrapper[4735]: I0215 20:31:15.872577 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" Feb 15 20:31:15 crc kubenswrapper[4735]: I0215 20:31:15.935620 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" podStartSLOduration=2.441713748 podStartE2EDuration="6.935593248s" podCreationTimestamp="2026-02-15 20:31:09 +0000 UTC" firstStartedPulling="2026-02-15 20:31:10.464134117 +0000 UTC m=+878.330149740" lastFinishedPulling="2026-02-15 20:31:14.958013577 +0000 UTC m=+882.824029240" observedRunningTime="2026-02-15 20:31:15.928244107 +0000 UTC m=+883.794259770" watchObservedRunningTime="2026-02-15 20:31:15.935593248 +0000 UTC m=+883.801608911" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.106445 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kp859"] Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.145860 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp859"] Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.146063 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.210520 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-58498c75bf-2kzsp" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.306217 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-utilities\") pod \"redhat-marketplace-kp859\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.306287 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljsts\" (UniqueName: \"kubernetes.io/projected/f51ec299-84f1-4f8c-af95-7535818d83cc-kube-api-access-ljsts\") pod \"redhat-marketplace-kp859\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.306355 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-catalog-content\") pod \"redhat-marketplace-kp859\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.407823 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-catalog-content\") pod \"redhat-marketplace-kp859\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.407929 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-utilities\") pod \"redhat-marketplace-kp859\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.408012 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljsts\" (UniqueName: \"kubernetes.io/projected/f51ec299-84f1-4f8c-af95-7535818d83cc-kube-api-access-ljsts\") pod \"redhat-marketplace-kp859\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.408745 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-catalog-content\") pod \"redhat-marketplace-kp859\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.408868 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-utilities\") pod \"redhat-marketplace-kp859\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.434031 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljsts\" (UniqueName: \"kubernetes.io/projected/f51ec299-84f1-4f8c-af95-7535818d83cc-kube-api-access-ljsts\") pod \"redhat-marketplace-kp859\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.477674 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:20 crc kubenswrapper[4735]: I0215 20:31:20.934091 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp859"] Feb 15 20:31:21 crc kubenswrapper[4735]: I0215 20:31:21.912431 4735 generic.go:334] "Generic (PLEG): container finished" podID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerID="4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385" exitCode=0 Feb 15 20:31:21 crc kubenswrapper[4735]: I0215 20:31:21.912670 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp859" event={"ID":"f51ec299-84f1-4f8c-af95-7535818d83cc","Type":"ContainerDied","Data":"4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385"} Feb 15 20:31:21 crc kubenswrapper[4735]: I0215 20:31:21.912894 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp859" event={"ID":"f51ec299-84f1-4f8c-af95-7535818d83cc","Type":"ContainerStarted","Data":"4ce60afd5c346cc2eeffda8ca00e938884d9c025b76919d0e3baeff8d12b745c"} Feb 15 20:31:22 crc kubenswrapper[4735]: I0215 20:31:22.920768 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp859" event={"ID":"f51ec299-84f1-4f8c-af95-7535818d83cc","Type":"ContainerStarted","Data":"e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe"} Feb 15 20:31:23 crc kubenswrapper[4735]: I0215 20:31:23.934904 4735 generic.go:334] "Generic (PLEG): container finished" podID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerID="e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe" exitCode=0 Feb 15 20:31:23 crc kubenswrapper[4735]: I0215 20:31:23.935000 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp859" event={"ID":"f51ec299-84f1-4f8c-af95-7535818d83cc","Type":"ContainerDied","Data":"e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe"} Feb 15 20:31:24 crc kubenswrapper[4735]: I0215 20:31:24.956483 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp859" event={"ID":"f51ec299-84f1-4f8c-af95-7535818d83cc","Type":"ContainerStarted","Data":"49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca"} Feb 15 20:31:24 crc kubenswrapper[4735]: I0215 20:31:24.979407 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kp859" podStartSLOduration=2.542504946 podStartE2EDuration="4.979391774s" podCreationTimestamp="2026-02-15 20:31:20 +0000 UTC" firstStartedPulling="2026-02-15 20:31:21.914290632 +0000 UTC m=+889.780306265" lastFinishedPulling="2026-02-15 20:31:24.35117747 +0000 UTC m=+892.217193093" observedRunningTime="2026-02-15 20:31:24.978352366 +0000 UTC m=+892.844367989" watchObservedRunningTime="2026-02-15 20:31:24.979391774 +0000 UTC m=+892.845407397" Feb 15 20:31:30 crc kubenswrapper[4735]: I0215 20:31:30.478033 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:30 crc kubenswrapper[4735]: I0215 20:31:30.478319 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:30 crc kubenswrapper[4735]: I0215 20:31:30.524770 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:31 crc kubenswrapper[4735]: I0215 20:31:31.030238 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:31 crc kubenswrapper[4735]: I0215 20:31:31.075288 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp859"] Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.000506 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kp859" podUID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerName="registry-server" containerID="cri-o://49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca" gracePeriod=2 Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.593132 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.682397 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-utilities\") pod \"f51ec299-84f1-4f8c-af95-7535818d83cc\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.682458 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljsts\" (UniqueName: \"kubernetes.io/projected/f51ec299-84f1-4f8c-af95-7535818d83cc-kube-api-access-ljsts\") pod \"f51ec299-84f1-4f8c-af95-7535818d83cc\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.682496 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-catalog-content\") pod \"f51ec299-84f1-4f8c-af95-7535818d83cc\" (UID: \"f51ec299-84f1-4f8c-af95-7535818d83cc\") " Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.702785 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f51ec299-84f1-4f8c-af95-7535818d83cc-kube-api-access-ljsts" (OuterVolumeSpecName: "kube-api-access-ljsts") pod "f51ec299-84f1-4f8c-af95-7535818d83cc" (UID: "f51ec299-84f1-4f8c-af95-7535818d83cc"). InnerVolumeSpecName "kube-api-access-ljsts". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.707082 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-utilities" (OuterVolumeSpecName: "utilities") pod "f51ec299-84f1-4f8c-af95-7535818d83cc" (UID: "f51ec299-84f1-4f8c-af95-7535818d83cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.714208 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f51ec299-84f1-4f8c-af95-7535818d83cc" (UID: "f51ec299-84f1-4f8c-af95-7535818d83cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.784213 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.784242 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljsts\" (UniqueName: \"kubernetes.io/projected/f51ec299-84f1-4f8c-af95-7535818d83cc-kube-api-access-ljsts\") on node \"crc\" DevicePath \"\"" Feb 15 20:31:33 crc kubenswrapper[4735]: I0215 20:31:33.784254 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f51ec299-84f1-4f8c-af95-7535818d83cc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.007399 4735 generic.go:334] "Generic (PLEG): container finished" podID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerID="49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca" exitCode=0 Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.007436 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp859" event={"ID":"f51ec299-84f1-4f8c-af95-7535818d83cc","Type":"ContainerDied","Data":"49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca"} Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.007469 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp859" event={"ID":"f51ec299-84f1-4f8c-af95-7535818d83cc","Type":"ContainerDied","Data":"4ce60afd5c346cc2eeffda8ca00e938884d9c025b76919d0e3baeff8d12b745c"} Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.007487 4735 scope.go:117] "RemoveContainer" containerID="49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.007499 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp859" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.020868 4735 scope.go:117] "RemoveContainer" containerID="e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.047059 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp859"] Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.050875 4735 scope.go:117] "RemoveContainer" containerID="4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.054312 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp859"] Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.067112 4735 scope.go:117] "RemoveContainer" containerID="49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca" Feb 15 20:31:34 crc kubenswrapper[4735]: E0215 20:31:34.067591 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca\": container with ID starting with 49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca not found: ID does not exist" containerID="49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.067637 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca"} err="failed to get container status \"49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca\": rpc error: code = NotFound desc = could not find container \"49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca\": container with ID starting with 49752379951bd1f8e0880b07bb5bd4393623081fc633fdebc4957b144aee5fca not found: ID does not exist" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.067677 4735 scope.go:117] "RemoveContainer" containerID="e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe" Feb 15 20:31:34 crc kubenswrapper[4735]: E0215 20:31:34.068193 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe\": container with ID starting with e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe not found: ID does not exist" containerID="e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.068210 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe"} err="failed to get container status \"e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe\": rpc error: code = NotFound desc = could not find container \"e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe\": container with ID starting with e3772fe99c1e6beaaee934ff09fdad7d9ea342782717145ca2a464f788000dbe not found: ID does not exist" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.068222 4735 scope.go:117] "RemoveContainer" containerID="4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385" Feb 15 20:31:34 crc kubenswrapper[4735]: E0215 20:31:34.068446 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385\": container with ID starting with 4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385 not found: ID does not exist" containerID="4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.068467 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385"} err="failed to get container status \"4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385\": rpc error: code = NotFound desc = could not find container \"4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385\": container with ID starting with 4e6725548a97ed35d62b0c3fab4d3d7cdf6fb8b5633465b838f1d1bfd0928385 not found: ID does not exist" Feb 15 20:31:34 crc kubenswrapper[4735]: I0215 20:31:34.893559 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f51ec299-84f1-4f8c-af95-7535818d83cc" path="/var/lib/kubelet/pods/f51ec299-84f1-4f8c-af95-7535818d83cc/volumes" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.171745 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f"] Feb 15 20:31:38 crc kubenswrapper[4735]: E0215 20:31:38.172608 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerName="extract-content" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.172624 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerName="extract-content" Feb 15 20:31:38 crc kubenswrapper[4735]: E0215 20:31:38.172647 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerName="extract-utilities" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.172655 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerName="extract-utilities" Feb 15 20:31:38 crc kubenswrapper[4735]: E0215 20:31:38.172672 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerName="registry-server" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.172683 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerName="registry-server" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.174023 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f51ec299-84f1-4f8c-af95-7535818d83cc" containerName="registry-server" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.174509 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.177590 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qfddd" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.188225 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.202482 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.203296 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.209518 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-8mqtq" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.217472 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.222898 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.223634 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.226936 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-nb292" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.240164 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.240958 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.248459 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pwb64" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.257085 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt64x\" (UniqueName: \"kubernetes.io/projected/98e57a69-0860-468e-898c-9ec42836d3d4-kube-api-access-mt64x\") pod \"cinder-operator-controller-manager-5d946d989d-v6nvk\" (UID: \"98e57a69-0860-468e-898c-9ec42836d3d4\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.257184 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d26ws\" (UniqueName: \"kubernetes.io/projected/99564ffe-1d8b-44b1-a625-4445d624a98f-kube-api-access-d26ws\") pod \"designate-operator-controller-manager-6d8bf5c495-vddvx\" (UID: \"99564ffe-1d8b-44b1-a625-4445d624a98f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.257223 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk97n\" (UniqueName: \"kubernetes.io/projected/b9be4a5f-fbec-4e52-bb00-cbc08576e1fd-kube-api-access-nk97n\") pod \"barbican-operator-controller-manager-868647ff47-9hc4f\" (UID: \"b9be4a5f-fbec-4e52-bb00-cbc08576e1fd\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.259724 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.267428 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.280127 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.282572 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.289721 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-85fkl" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.308289 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.314730 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.328928 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-gk27r" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.360785 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57ckg\" (UniqueName: \"kubernetes.io/projected/92aeecb8-b9fa-4db8-916e-f2dd800d2da3-kube-api-access-57ckg\") pod \"glance-operator-controller-manager-77987464f4-gvnkh\" (UID: \"92aeecb8-b9fa-4db8-916e-f2dd800d2da3\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.360839 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt64x\" (UniqueName: \"kubernetes.io/projected/98e57a69-0860-468e-898c-9ec42836d3d4-kube-api-access-mt64x\") pod \"cinder-operator-controller-manager-5d946d989d-v6nvk\" (UID: \"98e57a69-0860-468e-898c-9ec42836d3d4\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.360872 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltqgf\" (UniqueName: \"kubernetes.io/projected/ab28d8d3-58f1-460c-a0ed-75f61ebc2c52-kube-api-access-ltqgf\") pod \"heat-operator-controller-manager-69f49c598c-v4wpf\" (UID: \"ab28d8d3-58f1-460c-a0ed-75f61ebc2c52\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.360924 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d26ws\" (UniqueName: \"kubernetes.io/projected/99564ffe-1d8b-44b1-a625-4445d624a98f-kube-api-access-d26ws\") pod \"designate-operator-controller-manager-6d8bf5c495-vddvx\" (UID: \"99564ffe-1d8b-44b1-a625-4445d624a98f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.360941 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk97n\" (UniqueName: \"kubernetes.io/projected/b9be4a5f-fbec-4e52-bb00-cbc08576e1fd-kube-api-access-nk97n\") pod \"barbican-operator-controller-manager-868647ff47-9hc4f\" (UID: \"b9be4a5f-fbec-4e52-bb00-cbc08576e1fd\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.360979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k4sr\" (UniqueName: \"kubernetes.io/projected/31865132-118d-4f4c-bf0b-44ab13060882-kube-api-access-6k4sr\") pod \"horizon-operator-controller-manager-5b9b8895d5-fslrr\" (UID: \"31865132-118d-4f4c-bf0b-44ab13060882\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.375011 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.386420 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7676946c7c-5w649"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.387235 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.403025 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-npzdr" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.403088 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.416594 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt64x\" (UniqueName: \"kubernetes.io/projected/98e57a69-0860-468e-898c-9ec42836d3d4-kube-api-access-mt64x\") pod \"cinder-operator-controller-manager-5d946d989d-v6nvk\" (UID: \"98e57a69-0860-468e-898c-9ec42836d3d4\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.417649 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk97n\" (UniqueName: \"kubernetes.io/projected/b9be4a5f-fbec-4e52-bb00-cbc08576e1fd-kube-api-access-nk97n\") pod \"barbican-operator-controller-manager-868647ff47-9hc4f\" (UID: \"b9be4a5f-fbec-4e52-bb00-cbc08576e1fd\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.428492 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d26ws\" (UniqueName: \"kubernetes.io/projected/99564ffe-1d8b-44b1-a625-4445d624a98f-kube-api-access-d26ws\") pod \"designate-operator-controller-manager-6d8bf5c495-vddvx\" (UID: \"99564ffe-1d8b-44b1-a625-4445d624a98f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.438006 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.438800 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.442967 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-s6cfx" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.445223 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.446053 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.449134 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rbdx5" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.456470 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.462732 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkf6g\" (UniqueName: \"kubernetes.io/projected/79556e2d-fb3e-4a41-a85c-38005174ba50-kube-api-access-zkf6g\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.463976 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k4sr\" (UniqueName: \"kubernetes.io/projected/31865132-118d-4f4c-bf0b-44ab13060882-kube-api-access-6k4sr\") pod \"horizon-operator-controller-manager-5b9b8895d5-fslrr\" (UID: \"31865132-118d-4f4c-bf0b-44ab13060882\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.464032 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m9m4\" (UniqueName: \"kubernetes.io/projected/2b6ae856-4502-44cd-ad23-21577792b6a0-kube-api-access-4m9m4\") pod \"ironic-operator-controller-manager-554564d7fc-rbhc6\" (UID: \"2b6ae856-4502-44cd-ad23-21577792b6a0\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.464062 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57ckg\" (UniqueName: \"kubernetes.io/projected/92aeecb8-b9fa-4db8-916e-f2dd800d2da3-kube-api-access-57ckg\") pod \"glance-operator-controller-manager-77987464f4-gvnkh\" (UID: \"92aeecb8-b9fa-4db8-916e-f2dd800d2da3\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.470122 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.470203 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltqgf\" (UniqueName: \"kubernetes.io/projected/ab28d8d3-58f1-460c-a0ed-75f61ebc2c52-kube-api-access-ltqgf\") pod \"heat-operator-controller-manager-69f49c598c-v4wpf\" (UID: \"ab28d8d3-58f1-460c-a0ed-75f61ebc2c52\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.474557 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.479225 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.489036 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.490058 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.496316 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-g75g4" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.499292 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.500400 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.507167 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2z85k" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.509148 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.520559 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57ckg\" (UniqueName: \"kubernetes.io/projected/92aeecb8-b9fa-4db8-916e-f2dd800d2da3-kube-api-access-57ckg\") pod \"glance-operator-controller-manager-77987464f4-gvnkh\" (UID: \"92aeecb8-b9fa-4db8-916e-f2dd800d2da3\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.521013 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltqgf\" (UniqueName: \"kubernetes.io/projected/ab28d8d3-58f1-460c-a0ed-75f61ebc2c52-kube-api-access-ltqgf\") pod \"heat-operator-controller-manager-69f49c598c-v4wpf\" (UID: \"ab28d8d3-58f1-460c-a0ed-75f61ebc2c52\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.524902 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.526594 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k4sr\" (UniqueName: \"kubernetes.io/projected/31865132-118d-4f4c-bf0b-44ab13060882-kube-api-access-6k4sr\") pod \"horizon-operator-controller-manager-5b9b8895d5-fslrr\" (UID: \"31865132-118d-4f4c-bf0b-44ab13060882\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.550445 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.551205 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.552293 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.561506 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-qg2k6" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.562143 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.568199 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7676946c7c-5w649"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.578082 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cks2\" (UniqueName: \"kubernetes.io/projected/1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0-kube-api-access-5cks2\") pod \"manila-operator-controller-manager-54f6768c69-mh89r\" (UID: \"1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.578142 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m9m4\" (UniqueName: \"kubernetes.io/projected/2b6ae856-4502-44cd-ad23-21577792b6a0-kube-api-access-4m9m4\") pod \"ironic-operator-controller-manager-554564d7fc-rbhc6\" (UID: \"2b6ae856-4502-44cd-ad23-21577792b6a0\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.578174 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnzrl\" (UniqueName: \"kubernetes.io/projected/79e82794-8193-4eba-af98-f6e0ae43274b-kube-api-access-dnzrl\") pod \"mariadb-operator-controller-manager-6994f66f48-jwrfj\" (UID: \"79e82794-8193-4eba-af98-f6e0ae43274b\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.578197 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvj9c\" (UniqueName: \"kubernetes.io/projected/1094ce7c-b104-40d4-8e0f-3ee695bfd6cd-kube-api-access-bvj9c\") pod \"neutron-operator-controller-manager-64ddbf8bb-4hlsd\" (UID: \"1094ce7c-b104-40d4-8e0f-3ee695bfd6cd\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.578217 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.578250 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8f77\" (UniqueName: \"kubernetes.io/projected/8802a6f8-206d-473d-854c-77e3c74389b5-kube-api-access-r8f77\") pod \"keystone-operator-controller-manager-b4d948c87-ghs7m\" (UID: \"8802a6f8-206d-473d-854c-77e3c74389b5\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.578270 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkf6g\" (UniqueName: \"kubernetes.io/projected/79556e2d-fb3e-4a41-a85c-38005174ba50-kube-api-access-zkf6g\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:38 crc kubenswrapper[4735]: E0215 20:31:38.578670 4735 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:38 crc kubenswrapper[4735]: E0215 20:31:38.578707 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert podName:79556e2d-fb3e-4a41-a85c-38005174ba50 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:39.078693608 +0000 UTC m=+906.944709231 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert") pod "infra-operator-controller-manager-7676946c7c-5w649" (UID: "79556e2d-fb3e-4a41-a85c-38005174ba50") : secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.589990 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.609488 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.610262 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.619350 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-jvwlx" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.619937 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.622347 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkf6g\" (UniqueName: \"kubernetes.io/projected/79556e2d-fb3e-4a41-a85c-38005174ba50-kube-api-access-zkf6g\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.630703 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.632483 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m9m4\" (UniqueName: \"kubernetes.io/projected/2b6ae856-4502-44cd-ad23-21577792b6a0-kube-api-access-4m9m4\") pod \"ironic-operator-controller-manager-554564d7fc-rbhc6\" (UID: \"2b6ae856-4502-44cd-ad23-21577792b6a0\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.665018 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.666485 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.678935 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8f77\" (UniqueName: \"kubernetes.io/projected/8802a6f8-206d-473d-854c-77e3c74389b5-kube-api-access-r8f77\") pod \"keystone-operator-controller-manager-b4d948c87-ghs7m\" (UID: \"8802a6f8-206d-473d-854c-77e3c74389b5\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.678986 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb4df\" (UniqueName: \"kubernetes.io/projected/d8c734e9-fd8c-459d-982f-53b0a706d0a0-kube-api-access-fb4df\") pod \"nova-operator-controller-manager-567668f5cf-t9nc4\" (UID: \"d8c734e9-fd8c-459d-982f-53b0a706d0a0\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.679019 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cks2\" (UniqueName: \"kubernetes.io/projected/1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0-kube-api-access-5cks2\") pod \"manila-operator-controller-manager-54f6768c69-mh89r\" (UID: \"1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.679066 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnzrl\" (UniqueName: \"kubernetes.io/projected/79e82794-8193-4eba-af98-f6e0ae43274b-kube-api-access-dnzrl\") pod \"mariadb-operator-controller-manager-6994f66f48-jwrfj\" (UID: \"79e82794-8193-4eba-af98-f6e0ae43274b\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.679090 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvj9c\" (UniqueName: \"kubernetes.io/projected/1094ce7c-b104-40d4-8e0f-3ee695bfd6cd-kube-api-access-bvj9c\") pod \"neutron-operator-controller-manager-64ddbf8bb-4hlsd\" (UID: \"1094ce7c-b104-40d4-8e0f-3ee695bfd6cd\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.686850 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.691260 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.698107 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.705603 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-kppb6" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.722036 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.722838 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.728790 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-ljmzh" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.744482 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnzrl\" (UniqueName: \"kubernetes.io/projected/79e82794-8193-4eba-af98-f6e0ae43274b-kube-api-access-dnzrl\") pod \"mariadb-operator-controller-manager-6994f66f48-jwrfj\" (UID: \"79e82794-8193-4eba-af98-f6e0ae43274b\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.744912 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cks2\" (UniqueName: \"kubernetes.io/projected/1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0-kube-api-access-5cks2\") pod \"manila-operator-controller-manager-54f6768c69-mh89r\" (UID: \"1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.745752 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvj9c\" (UniqueName: \"kubernetes.io/projected/1094ce7c-b104-40d4-8e0f-3ee695bfd6cd-kube-api-access-bvj9c\") pod \"neutron-operator-controller-manager-64ddbf8bb-4hlsd\" (UID: \"1094ce7c-b104-40d4-8e0f-3ee695bfd6cd\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.749629 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.750366 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.751638 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8f77\" (UniqueName: \"kubernetes.io/projected/8802a6f8-206d-473d-854c-77e3c74389b5-kube-api-access-r8f77\") pod \"keystone-operator-controller-manager-b4d948c87-ghs7m\" (UID: \"8802a6f8-206d-473d-854c-77e3c74389b5\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.755651 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.755929 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-ppd7l" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.769807 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.776388 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.781879 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb4df\" (UniqueName: \"kubernetes.io/projected/d8c734e9-fd8c-459d-982f-53b0a706d0a0-kube-api-access-fb4df\") pod \"nova-operator-controller-manager-567668f5cf-t9nc4\" (UID: \"d8c734e9-fd8c-459d-982f-53b0a706d0a0\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.781971 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhzlz\" (UniqueName: \"kubernetes.io/projected/9549fdaf-fe4c-4da2-92db-db91beb95450-kube-api-access-mhzlz\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.782016 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.782048 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvcxg\" (UniqueName: \"kubernetes.io/projected/c7d8cf5f-13c8-4857-96d2-0da2c88ed221-kube-api-access-dvcxg\") pod \"octavia-operator-controller-manager-69f8888797-w8vz9\" (UID: \"c7d8cf5f-13c8-4857-96d2-0da2c88ed221\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.782081 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v82x\" (UniqueName: \"kubernetes.io/projected/1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea-kube-api-access-6v82x\") pod \"ovn-operator-controller-manager-d44cf6b75-f8s72\" (UID: \"1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.805586 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.811997 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.821267 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb4df\" (UniqueName: \"kubernetes.io/projected/d8c734e9-fd8c-459d-982f-53b0a706d0a0-kube-api-access-fb4df\") pod \"nova-operator-controller-manager-567668f5cf-t9nc4\" (UID: \"d8c734e9-fd8c-459d-982f-53b0a706d0a0\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.863582 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.865329 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.871841 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-jpvgf" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.872372 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.883698 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhzlz\" (UniqueName: \"kubernetes.io/projected/9549fdaf-fe4c-4da2-92db-db91beb95450-kube-api-access-mhzlz\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.883757 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.883781 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvcxg\" (UniqueName: \"kubernetes.io/projected/c7d8cf5f-13c8-4857-96d2-0da2c88ed221-kube-api-access-dvcxg\") pod \"octavia-operator-controller-manager-69f8888797-w8vz9\" (UID: \"c7d8cf5f-13c8-4857-96d2-0da2c88ed221\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.883816 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v82x\" (UniqueName: \"kubernetes.io/projected/1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea-kube-api-access-6v82x\") pod \"ovn-operator-controller-manager-d44cf6b75-f8s72\" (UID: \"1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" Feb 15 20:31:38 crc kubenswrapper[4735]: E0215 20:31:38.884258 4735 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:38 crc kubenswrapper[4735]: E0215 20:31:38.884295 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert podName:9549fdaf-fe4c-4da2-92db-db91beb95450 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:39.384282942 +0000 UTC m=+907.250298565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" (UID: "9549fdaf-fe4c-4da2-92db-db91beb95450") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.947046 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvcxg\" (UniqueName: \"kubernetes.io/projected/c7d8cf5f-13c8-4857-96d2-0da2c88ed221-kube-api-access-dvcxg\") pod \"octavia-operator-controller-manager-69f8888797-w8vz9\" (UID: \"c7d8cf5f-13c8-4857-96d2-0da2c88ed221\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.947847 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v82x\" (UniqueName: \"kubernetes.io/projected/1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea-kube-api-access-6v82x\") pod \"ovn-operator-controller-manager-d44cf6b75-f8s72\" (UID: \"1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.966626 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhzlz\" (UniqueName: \"kubernetes.io/projected/9549fdaf-fe4c-4da2-92db-db91beb95450-kube-api-access-mhzlz\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.968278 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.973700 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n"] Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.982425 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.982564 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" Feb 15 20:31:38 crc kubenswrapper[4735]: I0215 20:31:38.986326 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-552rp\" (UniqueName: \"kubernetes.io/projected/d8bf1d9d-df22-4b17-91b7-833052b5a337-kube-api-access-552rp\") pod \"placement-operator-controller-manager-8497b45c89-9trcb\" (UID: \"d8bf1d9d-df22-4b17-91b7-833052b5a337\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:38.989548 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:38.989628 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.012466 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.014257 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.016114 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.019522 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-xg6z8" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.034819 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-d58rt" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.034976 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.040623 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-r6bz7"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.042752 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.044822 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.049273 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-kmhpv" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.049580 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.052864 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-r6bz7"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.077795 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.080004 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.080701 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.084537 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-xknd6" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.085239 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.106876 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s82q\" (UniqueName: \"kubernetes.io/projected/3fa6a777-e009-4e48-b05d-ac9b34acd21c-kube-api-access-5s82q\") pod \"test-operator-controller-manager-7866795846-r6bz7\" (UID: \"3fa6a777-e009-4e48-b05d-ac9b34acd21c\") " pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.106958 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-552rp\" (UniqueName: \"kubernetes.io/projected/d8bf1d9d-df22-4b17-91b7-833052b5a337-kube-api-access-552rp\") pod \"placement-operator-controller-manager-8497b45c89-9trcb\" (UID: \"d8bf1d9d-df22-4b17-91b7-833052b5a337\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.106983 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jntql\" (UniqueName: \"kubernetes.io/projected/4fb2a283-35ca-4391-8e59-79a8fa7bb4f0-kube-api-access-jntql\") pod \"telemetry-operator-controller-manager-7f45b4ff68-lfj2m\" (UID: \"4fb2a283-35ca-4391-8e59-79a8fa7bb4f0\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.107006 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqsdh\" (UniqueName: \"kubernetes.io/projected/14de9f25-fdbd-4d44-9f2c-440b256c4faf-kube-api-access-hqsdh\") pod \"swift-operator-controller-manager-68f46476f-vbr6n\" (UID: \"14de9f25-fdbd-4d44-9f2c-440b256c4faf\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.109093 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.109266 4735 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.109341 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert podName:79556e2d-fb3e-4a41-a85c-38005174ba50 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:40.10930437 +0000 UTC m=+907.975319993 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert") pod "infra-operator-controller-manager-7676946c7c-5w649" (UID: "79556e2d-fb3e-4a41-a85c-38005174ba50") : secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.159330 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rn68p"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.160759 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.184247 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rn68p"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.194511 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-552rp\" (UniqueName: \"kubernetes.io/projected/d8bf1d9d-df22-4b17-91b7-833052b5a337-kube-api-access-552rp\") pod \"placement-operator-controller-manager-8497b45c89-9trcb\" (UID: \"d8bf1d9d-df22-4b17-91b7-833052b5a337\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.194574 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.195404 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.209339 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-2fsr4" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.209848 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.210036 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.211348 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-utilities\") pod \"community-operators-rn68p\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.211452 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jntql\" (UniqueName: \"kubernetes.io/projected/4fb2a283-35ca-4391-8e59-79a8fa7bb4f0-kube-api-access-jntql\") pod \"telemetry-operator-controller-manager-7f45b4ff68-lfj2m\" (UID: \"4fb2a283-35ca-4391-8e59-79a8fa7bb4f0\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.211525 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqsdh\" (UniqueName: \"kubernetes.io/projected/14de9f25-fdbd-4d44-9f2c-440b256c4faf-kube-api-access-hqsdh\") pod \"swift-operator-controller-manager-68f46476f-vbr6n\" (UID: \"14de9f25-fdbd-4d44-9f2c-440b256c4faf\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.211589 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtvgs\" (UniqueName: \"kubernetes.io/projected/147bc244-16d1-4e31-8ff4-f3b129803fd5-kube-api-access-wtvgs\") pod \"community-operators-rn68p\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.211657 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-catalog-content\") pod \"community-operators-rn68p\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.211788 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s82q\" (UniqueName: \"kubernetes.io/projected/3fa6a777-e009-4e48-b05d-ac9b34acd21c-kube-api-access-5s82q\") pod \"test-operator-controller-manager-7866795846-r6bz7\" (UID: \"3fa6a777-e009-4e48-b05d-ac9b34acd21c\") " pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.211856 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6wt2\" (UniqueName: \"kubernetes.io/projected/26e7f3b0-6f34-491e-917a-e545d320ee89-kube-api-access-z6wt2\") pod \"watcher-operator-controller-manager-5db88f68c-d65l6\" (UID: \"26e7f3b0-6f34-491e-917a-e545d320ee89\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.216320 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.240355 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqsdh\" (UniqueName: \"kubernetes.io/projected/14de9f25-fdbd-4d44-9f2c-440b256c4faf-kube-api-access-hqsdh\") pod \"swift-operator-controller-manager-68f46476f-vbr6n\" (UID: \"14de9f25-fdbd-4d44-9f2c-440b256c4faf\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.240412 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.241244 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.244615 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.247223 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-q54gf" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.250392 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.252593 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s82q\" (UniqueName: \"kubernetes.io/projected/3fa6a777-e009-4e48-b05d-ac9b34acd21c-kube-api-access-5s82q\") pod \"test-operator-controller-manager-7866795846-r6bz7\" (UID: \"3fa6a777-e009-4e48-b05d-ac9b34acd21c\") " pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.253491 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jntql\" (UniqueName: \"kubernetes.io/projected/4fb2a283-35ca-4391-8e59-79a8fa7bb4f0-kube-api-access-jntql\") pod \"telemetry-operator-controller-manager-7f45b4ff68-lfj2m\" (UID: \"4fb2a283-35ca-4391-8e59-79a8fa7bb4f0\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.314202 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.314265 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.314287 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6wt2\" (UniqueName: \"kubernetes.io/projected/26e7f3b0-6f34-491e-917a-e545d320ee89-kube-api-access-z6wt2\") pod \"watcher-operator-controller-manager-5db88f68c-d65l6\" (UID: \"26e7f3b0-6f34-491e-917a-e545d320ee89\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.314316 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l74qv\" (UniqueName: \"kubernetes.io/projected/6fad7194-6e10-4dc9-ac3c-241722357e55-kube-api-access-l74qv\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.314349 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-utilities\") pod \"community-operators-rn68p\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.314374 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqlhk\" (UniqueName: \"kubernetes.io/projected/e57b4eea-5c8f-4caa-9644-0994108974da-kube-api-access-pqlhk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9z8vz\" (UID: \"e57b4eea-5c8f-4caa-9644-0994108974da\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.314400 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtvgs\" (UniqueName: \"kubernetes.io/projected/147bc244-16d1-4e31-8ff4-f3b129803fd5-kube-api-access-wtvgs\") pod \"community-operators-rn68p\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.314421 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-catalog-content\") pod \"community-operators-rn68p\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.315894 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-catalog-content\") pod \"community-operators-rn68p\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.316408 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-utilities\") pod \"community-operators-rn68p\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.350224 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtvgs\" (UniqueName: \"kubernetes.io/projected/147bc244-16d1-4e31-8ff4-f3b129803fd5-kube-api-access-wtvgs\") pod \"community-operators-rn68p\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.359108 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6wt2\" (UniqueName: \"kubernetes.io/projected/26e7f3b0-6f34-491e-917a-e545d320ee89-kube-api-access-z6wt2\") pod \"watcher-operator-controller-manager-5db88f68c-d65l6\" (UID: \"26e7f3b0-6f34-491e-917a-e545d320ee89\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.386364 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.393579 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.417425 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.417466 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.417511 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.417541 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l74qv\" (UniqueName: \"kubernetes.io/projected/6fad7194-6e10-4dc9-ac3c-241722357e55-kube-api-access-l74qv\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.418124 4735 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.418201 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:39.918183451 +0000 UTC m=+907.784199074 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "webhook-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.418399 4735 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.418424 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert podName:9549fdaf-fe4c-4da2-92db-db91beb95450 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:40.418417737 +0000 UTC m=+908.284433360 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" (UID: "9549fdaf-fe4c-4da2-92db-db91beb95450") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.418620 4735 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.418672 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:39.918656593 +0000 UTC m=+907.784672216 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "metrics-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.425607 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.432034 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqlhk\" (UniqueName: \"kubernetes.io/projected/e57b4eea-5c8f-4caa-9644-0994108974da-kube-api-access-pqlhk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9z8vz\" (UID: \"e57b4eea-5c8f-4caa-9644-0994108974da\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.474304 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqlhk\" (UniqueName: \"kubernetes.io/projected/e57b4eea-5c8f-4caa-9644-0994108974da-kube-api-access-pqlhk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9z8vz\" (UID: \"e57b4eea-5c8f-4caa-9644-0994108974da\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.474719 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l74qv\" (UniqueName: \"kubernetes.io/projected/6fad7194-6e10-4dc9-ac3c-241722357e55-kube-api-access-l74qv\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.560242 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.649481 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.715267 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" Feb 15 20:31:39 crc kubenswrapper[4735]: W0215 20:31:39.903666 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9be4a5f_fbec_4e52_bb00_cbc08576e1fd.slice/crio-818263f089c3af83cb8c0868f1c95f8eb48eeaf9cc43538df29295d9c49e4159 WatchSource:0}: Error finding container 818263f089c3af83cb8c0868f1c95f8eb48eeaf9cc43538df29295d9c49e4159: Status 404 returned error can't find the container with id 818263f089c3af83cb8c0868f1c95f8eb48eeaf9cc43538df29295d9c49e4159 Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.904927 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.915758 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr"] Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.942857 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: I0215 20:31:39.942925 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.943193 4735 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.943237 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:40.943223728 +0000 UTC m=+908.809239351 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "metrics-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.944405 4735 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: E0215 20:31:39.944468 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:40.944451031 +0000 UTC m=+908.810466654 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "webhook-server-cert" not found Feb 15 20:31:39 crc kubenswrapper[4735]: W0215 20:31:39.946172 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31865132_118d_4f4c_bf0b_44ab13060882.slice/crio-28fafeb78f5d178b177093cf55558b199df5f6570780dae95d265b50cc76c838 WatchSource:0}: Error finding container 28fafeb78f5d178b177093cf55558b199df5f6570780dae95d265b50cc76c838: Status 404 returned error can't find the container with id 28fafeb78f5d178b177093cf55558b199df5f6570780dae95d265b50cc76c838 Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.054003 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" event={"ID":"b9be4a5f-fbec-4e52-bb00-cbc08576e1fd","Type":"ContainerStarted","Data":"818263f089c3af83cb8c0868f1c95f8eb48eeaf9cc43538df29295d9c49e4159"} Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.054834 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" event={"ID":"31865132-118d-4f4c-bf0b-44ab13060882","Type":"ContainerStarted","Data":"28fafeb78f5d178b177093cf55558b199df5f6570780dae95d265b50cc76c838"} Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.089362 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh"] Feb 15 20:31:40 crc kubenswrapper[4735]: W0215 20:31:40.125850 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92aeecb8_b9fa_4db8_916e_f2dd800d2da3.slice/crio-b234a64c67ac2e6557327cd60942afba81fc283d172af87c383f278ec629ec7a WatchSource:0}: Error finding container b234a64c67ac2e6557327cd60942afba81fc283d172af87c383f278ec629ec7a: Status 404 returned error can't find the container with id b234a64c67ac2e6557327cd60942afba81fc283d172af87c383f278ec629ec7a Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.150669 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:40 crc kubenswrapper[4735]: E0215 20:31:40.150815 4735 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:40 crc kubenswrapper[4735]: E0215 20:31:40.150859 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert podName:79556e2d-fb3e-4a41-a85c-38005174ba50 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:42.150845293 +0000 UTC m=+910.016860916 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert") pod "infra-operator-controller-manager-7676946c7c-5w649" (UID: "79556e2d-fb3e-4a41-a85c-38005174ba50") : secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.155639 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx"] Feb 15 20:31:40 crc kubenswrapper[4735]: W0215 20:31:40.169879 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99564ffe_1d8b_44b1_a625_4445d624a98f.slice/crio-d1a6a97d3f514fe3505d4d2542659a5a9c1cbb8792366d2bde57b1701ac5f087 WatchSource:0}: Error finding container d1a6a97d3f514fe3505d4d2542659a5a9c1cbb8792366d2bde57b1701ac5f087: Status 404 returned error can't find the container with id d1a6a97d3f514fe3505d4d2542659a5a9c1cbb8792366d2bde57b1701ac5f087 Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.232625 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk"] Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.458574 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:40 crc kubenswrapper[4735]: E0215 20:31:40.458769 4735 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:40 crc kubenswrapper[4735]: E0215 20:31:40.458821 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert podName:9549fdaf-fe4c-4da2-92db-db91beb95450 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:42.458804629 +0000 UTC m=+910.324820252 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" (UID: "9549fdaf-fe4c-4da2-92db-db91beb95450") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.633414 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6"] Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.652600 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf"] Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.684249 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd"] Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.733804 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m"] Feb 15 20:31:40 crc kubenswrapper[4735]: W0215 20:31:40.740577 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b6ae856_4502_44cd_ad23_21577792b6a0.slice/crio-90205249672f35cf9f55382ad8e06ed611f451fb3146b29e7bbb2c7a2a16772e WatchSource:0}: Error finding container 90205249672f35cf9f55382ad8e06ed611f451fb3146b29e7bbb2c7a2a16772e: Status 404 returned error can't find the container with id 90205249672f35cf9f55382ad8e06ed611f451fb3146b29e7bbb2c7a2a16772e Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.787091 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4"] Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.822419 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9"] Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.866876 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m"] Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.913216 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-r6bz7"] Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.952390 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r"] Feb 15 20:31:40 crc kubenswrapper[4735]: W0215 20:31:40.968401 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fa6a777_e009_4e48_b05d_ac9b34acd21c.slice/crio-ff50eb14859dbc56c6001057d66fff775dbc370954d9b456dade61b831de3122 WatchSource:0}: Error finding container ff50eb14859dbc56c6001057d66fff775dbc370954d9b456dade61b831de3122: Status 404 returned error can't find the container with id ff50eb14859dbc56c6001057d66fff775dbc370954d9b456dade61b831de3122 Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.972714 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:40 crc kubenswrapper[4735]: I0215 20:31:40.972801 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:40 crc kubenswrapper[4735]: E0215 20:31:40.972963 4735 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 15 20:31:40 crc kubenswrapper[4735]: E0215 20:31:40.973012 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:42.972995392 +0000 UTC m=+910.839011015 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "metrics-server-cert" not found Feb 15 20:31:40 crc kubenswrapper[4735]: E0215 20:31:40.973256 4735 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 15 20:31:40 crc kubenswrapper[4735]: E0215 20:31:40.973348 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:42.97332952 +0000 UTC m=+910.839345143 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "webhook-server-cert" not found Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.009504 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rn68p"] Feb 15 20:31:41 crc kubenswrapper[4735]: W0215 20:31:41.043083 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod147bc244_16d1_4e31_8ff4_f3b129803fd5.slice/crio-b2fde80dba9d1ba124d35fe0293fea09655af2c0b0b31cd2249b637e5004479f WatchSource:0}: Error finding container b2fde80dba9d1ba124d35fe0293fea09655af2c0b0b31cd2249b637e5004479f: Status 404 returned error can't find the container with id b2fde80dba9d1ba124d35fe0293fea09655af2c0b0b31cd2249b637e5004479f Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.056006 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6"] Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.076011 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" event={"ID":"ab28d8d3-58f1-460c-a0ed-75f61ebc2c52","Type":"ContainerStarted","Data":"57539fc7ae19b08ae088be4f700218cee0a70fe34b0d72854980d2548af61f56"} Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.079248 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n"] Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.094429 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" event={"ID":"8802a6f8-206d-473d-854c-77e3c74389b5","Type":"ContainerStarted","Data":"0a44ae16ff3e56b9d6e1e4c6428876e3b88071b3f5ee0ea55754d5e956488ba8"} Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.106490 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb"] Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.114492 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" event={"ID":"4fb2a283-35ca-4391-8e59-79a8fa7bb4f0","Type":"ContainerStarted","Data":"d9e9de3c2f08a4f473d1d6ee275ccf59947111684573810d4ab3f165c65176ca"} Feb 15 20:31:41 crc kubenswrapper[4735]: W0215 20:31:41.114571 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26e7f3b0_6f34_491e_917a_e545d320ee89.slice/crio-d31ba823679eaf380483e6e56b7dc57f4bda5c380f8afbb45149129bdffc806a WatchSource:0}: Error finding container d31ba823679eaf380483e6e56b7dc57f4bda5c380f8afbb45149129bdffc806a: Status 404 returned error can't find the container with id d31ba823679eaf380483e6e56b7dc57f4bda5c380f8afbb45149129bdffc806a Feb 15 20:31:41 crc kubenswrapper[4735]: E0215 20:31:41.114658 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dnzrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6994f66f48-jwrfj_openstack-operators(79e82794-8193-4eba-af98-f6e0ae43274b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 15 20:31:41 crc kubenswrapper[4735]: E0215 20:31:41.117239 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" podUID="79e82794-8193-4eba-af98-f6e0ae43274b" Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.120994 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" event={"ID":"c7d8cf5f-13c8-4857-96d2-0da2c88ed221","Type":"ContainerStarted","Data":"33ed045dce7d09dee65a083a2884a7dc656e05c7f4c65b29b851ffddfd02863d"} Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.128026 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" event={"ID":"1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0","Type":"ContainerStarted","Data":"1475d351e234014160b20d7692118268c58abd28c0dc00d6c20b38fee989a548"} Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.133169 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj"] Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.144567 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz"] Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.149325 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" event={"ID":"3fa6a777-e009-4e48-b05d-ac9b34acd21c","Type":"ContainerStarted","Data":"ff50eb14859dbc56c6001057d66fff775dbc370954d9b456dade61b831de3122"} Feb 15 20:31:41 crc kubenswrapper[4735]: E0215 20:31:41.151249 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6v82x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-d44cf6b75-f8s72_openstack-operators(1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 15 20:31:41 crc kubenswrapper[4735]: E0215 20:31:41.151385 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-552rp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8497b45c89-9trcb_openstack-operators(d8bf1d9d-df22-4b17-91b7-833052b5a337): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 15 20:31:41 crc kubenswrapper[4735]: E0215 20:31:41.152379 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" podUID="1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea" Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.152519 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" event={"ID":"1094ce7c-b104-40d4-8e0f-3ee695bfd6cd","Type":"ContainerStarted","Data":"c1953ebfff1a4f3dd83cc7bb6f21ccaacdb3e5e9c6f10453e4faf97f40cc4eab"} Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.166858 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72"] Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.167028 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" event={"ID":"d8c734e9-fd8c-459d-982f-53b0a706d0a0","Type":"ContainerStarted","Data":"6cbbe3ef5d11b3ed584ca199a24e8cd221172638dcb6d8df3a68abf84f7a48d6"} Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.167124 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" event={"ID":"98e57a69-0860-468e-898c-9ec42836d3d4","Type":"ContainerStarted","Data":"fdeb67ba52a06fd3372d6d36d57a9e401114ffac8a8cc70418d7607ec8725b52"} Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.167212 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" event={"ID":"99564ffe-1d8b-44b1-a625-4445d624a98f","Type":"ContainerStarted","Data":"d1a6a97d3f514fe3505d4d2542659a5a9c1cbb8792366d2bde57b1701ac5f087"} Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.167301 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" event={"ID":"2b6ae856-4502-44cd-ad23-21577792b6a0","Type":"ContainerStarted","Data":"90205249672f35cf9f55382ad8e06ed611f451fb3146b29e7bbb2c7a2a16772e"} Feb 15 20:31:41 crc kubenswrapper[4735]: E0215 20:31:41.164734 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pqlhk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-9z8vz_openstack-operators(e57b4eea-5c8f-4caa-9644-0994108974da): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 15 20:31:41 crc kubenswrapper[4735]: E0215 20:31:41.152578 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" podUID="d8bf1d9d-df22-4b17-91b7-833052b5a337" Feb 15 20:31:41 crc kubenswrapper[4735]: E0215 20:31:41.169623 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" podUID="e57b4eea-5c8f-4caa-9644-0994108974da" Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.173293 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" event={"ID":"92aeecb8-b9fa-4db8-916e-f2dd800d2da3","Type":"ContainerStarted","Data":"b234a64c67ac2e6557327cd60942afba81fc283d172af87c383f278ec629ec7a"} Feb 15 20:31:41 crc kubenswrapper[4735]: I0215 20:31:41.177528 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn68p" event={"ID":"147bc244-16d1-4e31-8ff4-f3b129803fd5","Type":"ContainerStarted","Data":"b2fde80dba9d1ba124d35fe0293fea09655af2c0b0b31cd2249b637e5004479f"} Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.195988 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:42 crc kubenswrapper[4735]: E0215 20:31:42.196256 4735 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:42 crc kubenswrapper[4735]: E0215 20:31:42.196320 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert podName:79556e2d-fb3e-4a41-a85c-38005174ba50 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:46.196290434 +0000 UTC m=+914.062306057 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert") pod "infra-operator-controller-manager-7676946c7c-5w649" (UID: "79556e2d-fb3e-4a41-a85c-38005174ba50") : secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.215464 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" event={"ID":"79e82794-8193-4eba-af98-f6e0ae43274b","Type":"ContainerStarted","Data":"7ac8ccd403a64fd6449654bd9b53d388d0df7fa5f19ad9f1d12f7936d54c2e8e"} Feb 15 20:31:42 crc kubenswrapper[4735]: E0215 20:31:42.216807 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" podUID="79e82794-8193-4eba-af98-f6e0ae43274b" Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.217779 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" event={"ID":"14de9f25-fdbd-4d44-9f2c-440b256c4faf","Type":"ContainerStarted","Data":"e20995213ecaedd3dd0eb23da6d2589536824369cf5ac1ba71b6d2cf42c89256"} Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.235863 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" event={"ID":"d8bf1d9d-df22-4b17-91b7-833052b5a337","Type":"ContainerStarted","Data":"1f9474a97a0dcddd6a389fead0302dd08fbac65332a737bd155cde97b5eb198f"} Feb 15 20:31:42 crc kubenswrapper[4735]: E0215 20:31:42.246613 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" podUID="d8bf1d9d-df22-4b17-91b7-833052b5a337" Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.254441 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" event={"ID":"1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea","Type":"ContainerStarted","Data":"1650786d80a5b3a4be9dfae6033ce74543d6226bdfe698c9db4899860389a7bf"} Feb 15 20:31:42 crc kubenswrapper[4735]: E0215 20:31:42.259688 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" podUID="1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea" Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.293226 4735 generic.go:334] "Generic (PLEG): container finished" podID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerID="1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99" exitCode=0 Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.293390 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn68p" event={"ID":"147bc244-16d1-4e31-8ff4-f3b129803fd5","Type":"ContainerDied","Data":"1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99"} Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.297085 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" event={"ID":"26e7f3b0-6f34-491e-917a-e545d320ee89","Type":"ContainerStarted","Data":"d31ba823679eaf380483e6e56b7dc57f4bda5c380f8afbb45149129bdffc806a"} Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.320290 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" event={"ID":"e57b4eea-5c8f-4caa-9644-0994108974da","Type":"ContainerStarted","Data":"09ca6fa986b858af946d015c90b69353ce679417acca68676bc59ed59b43852d"} Feb 15 20:31:42 crc kubenswrapper[4735]: E0215 20:31:42.322296 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" podUID="e57b4eea-5c8f-4caa-9644-0994108974da" Feb 15 20:31:42 crc kubenswrapper[4735]: I0215 20:31:42.500254 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:42 crc kubenswrapper[4735]: E0215 20:31:42.500598 4735 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:42 crc kubenswrapper[4735]: E0215 20:31:42.500939 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert podName:9549fdaf-fe4c-4da2-92db-db91beb95450 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:46.500920808 +0000 UTC m=+914.366936431 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" (UID: "9549fdaf-fe4c-4da2-92db-db91beb95450") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:43 crc kubenswrapper[4735]: I0215 20:31:43.009830 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:43 crc kubenswrapper[4735]: I0215 20:31:43.009894 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:43 crc kubenswrapper[4735]: E0215 20:31:43.010076 4735 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 15 20:31:43 crc kubenswrapper[4735]: E0215 20:31:43.010125 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:47.01010558 +0000 UTC m=+914.876121203 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "metrics-server-cert" not found Feb 15 20:31:43 crc kubenswrapper[4735]: E0215 20:31:43.010169 4735 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 15 20:31:43 crc kubenswrapper[4735]: E0215 20:31:43.010193 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:47.010186142 +0000 UTC m=+914.876201765 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "webhook-server-cert" not found Feb 15 20:31:43 crc kubenswrapper[4735]: I0215 20:31:43.355166 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn68p" event={"ID":"147bc244-16d1-4e31-8ff4-f3b129803fd5","Type":"ContainerStarted","Data":"5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f"} Feb 15 20:31:43 crc kubenswrapper[4735]: E0215 20:31:43.359427 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" podUID="e57b4eea-5c8f-4caa-9644-0994108974da" Feb 15 20:31:43 crc kubenswrapper[4735]: E0215 20:31:43.361611 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" podUID="1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea" Feb 15 20:31:43 crc kubenswrapper[4735]: E0215 20:31:43.361880 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" podUID="d8bf1d9d-df22-4b17-91b7-833052b5a337" Feb 15 20:31:43 crc kubenswrapper[4735]: E0215 20:31:43.371624 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" podUID="79e82794-8193-4eba-af98-f6e0ae43274b" Feb 15 20:31:44 crc kubenswrapper[4735]: I0215 20:31:44.379928 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn68p" event={"ID":"147bc244-16d1-4e31-8ff4-f3b129803fd5","Type":"ContainerDied","Data":"5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f"} Feb 15 20:31:44 crc kubenswrapper[4735]: I0215 20:31:44.379997 4735 generic.go:334] "Generic (PLEG): container finished" podID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerID="5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f" exitCode=0 Feb 15 20:31:46 crc kubenswrapper[4735]: I0215 20:31:46.282036 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:46 crc kubenswrapper[4735]: E0215 20:31:46.282238 4735 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:46 crc kubenswrapper[4735]: E0215 20:31:46.282864 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert podName:79556e2d-fb3e-4a41-a85c-38005174ba50 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:54.282844462 +0000 UTC m=+922.148860085 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert") pod "infra-operator-controller-manager-7676946c7c-5w649" (UID: "79556e2d-fb3e-4a41-a85c-38005174ba50") : secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:46 crc kubenswrapper[4735]: I0215 20:31:46.585813 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:46 crc kubenswrapper[4735]: E0215 20:31:46.586000 4735 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:46 crc kubenswrapper[4735]: E0215 20:31:46.586285 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert podName:9549fdaf-fe4c-4da2-92db-db91beb95450 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:54.586269894 +0000 UTC m=+922.452285517 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" (UID: "9549fdaf-fe4c-4da2-92db-db91beb95450") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:47 crc kubenswrapper[4735]: I0215 20:31:47.092338 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:47 crc kubenswrapper[4735]: I0215 20:31:47.092401 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:47 crc kubenswrapper[4735]: E0215 20:31:47.092513 4735 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 15 20:31:47 crc kubenswrapper[4735]: E0215 20:31:47.092557 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:55.092543438 +0000 UTC m=+922.958559061 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "metrics-server-cert" not found Feb 15 20:31:47 crc kubenswrapper[4735]: E0215 20:31:47.092597 4735 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 15 20:31:47 crc kubenswrapper[4735]: E0215 20:31:47.092615 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:31:55.09260888 +0000 UTC m=+922.958624513 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "webhook-server-cert" not found Feb 15 20:31:49 crc kubenswrapper[4735]: I0215 20:31:49.680087 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:31:49 crc kubenswrapper[4735]: I0215 20:31:49.680148 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:31:54 crc kubenswrapper[4735]: I0215 20:31:54.300695 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:31:54 crc kubenswrapper[4735]: E0215 20:31:54.300879 4735 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:54 crc kubenswrapper[4735]: E0215 20:31:54.301399 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert podName:79556e2d-fb3e-4a41-a85c-38005174ba50 nodeName:}" failed. No retries permitted until 2026-02-15 20:32:10.301381243 +0000 UTC m=+938.167396866 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert") pod "infra-operator-controller-manager-7676946c7c-5w649" (UID: "79556e2d-fb3e-4a41-a85c-38005174ba50") : secret "infra-operator-webhook-server-cert" not found Feb 15 20:31:54 crc kubenswrapper[4735]: I0215 20:31:54.604753 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:31:54 crc kubenswrapper[4735]: E0215 20:31:54.604892 4735 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:54 crc kubenswrapper[4735]: E0215 20:31:54.605310 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert podName:9549fdaf-fe4c-4da2-92db-db91beb95450 nodeName:}" failed. No retries permitted until 2026-02-15 20:32:10.605287387 +0000 UTC m=+938.471303030 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" (UID: "9549fdaf-fe4c-4da2-92db-db91beb95450") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 15 20:31:55 crc kubenswrapper[4735]: I0215 20:31:55.113306 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:55 crc kubenswrapper[4735]: I0215 20:31:55.113424 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:31:55 crc kubenswrapper[4735]: E0215 20:31:55.113453 4735 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 15 20:31:55 crc kubenswrapper[4735]: E0215 20:31:55.113535 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:32:11.113517363 +0000 UTC m=+938.979532986 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "metrics-server-cert" not found Feb 15 20:31:55 crc kubenswrapper[4735]: E0215 20:31:55.113587 4735 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 15 20:31:55 crc kubenswrapper[4735]: E0215 20:31:55.113636 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs podName:6fad7194-6e10-4dc9-ac3c-241722357e55 nodeName:}" failed. No retries permitted until 2026-02-15 20:32:11.113621947 +0000 UTC m=+938.979637560 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs") pod "openstack-operator-controller-manager-79f9b74c8-4b426" (UID: "6fad7194-6e10-4dc9-ac3c-241722357e55") : secret "webhook-server-cert" not found Feb 15 20:32:00 crc kubenswrapper[4735]: E0215 20:32:00.141194 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da" Feb 15 20:32:00 crc kubenswrapper[4735]: E0215 20:32:00.142818 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6k4sr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5b9b8895d5-fslrr_openstack-operators(31865132-118d-4f4c-bf0b-44ab13060882): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:32:00 crc kubenswrapper[4735]: E0215 20:32:00.144108 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" podUID="31865132-118d-4f4c-bf0b-44ab13060882" Feb 15 20:32:00 crc kubenswrapper[4735]: E0215 20:32:00.504351 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" podUID="31865132-118d-4f4c-bf0b-44ab13060882" Feb 15 20:32:01 crc kubenswrapper[4735]: E0215 20:32:01.759189 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf" Feb 15 20:32:01 crc kubenswrapper[4735]: E0215 20:32:01.759829 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bvj9c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-64ddbf8bb-4hlsd_openstack-operators(1094ce7c-b104-40d4-8e0f-3ee695bfd6cd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:32:01 crc kubenswrapper[4735]: E0215 20:32:01.761199 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" podUID="1094ce7c-b104-40d4-8e0f-3ee695bfd6cd" Feb 15 20:32:02 crc kubenswrapper[4735]: E0215 20:32:02.517756 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" podUID="1094ce7c-b104-40d4-8e0f-3ee695bfd6cd" Feb 15 20:32:03 crc kubenswrapper[4735]: E0215 20:32:03.123767 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:90ad8fd8c1889b6be77925016532218eb6149d2c1c8535a5f9f1775c776fa6cc" Feb 15 20:32:03 crc kubenswrapper[4735]: E0215 20:32:03.124011 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:90ad8fd8c1889b6be77925016532218eb6149d2c1c8535a5f9f1775c776fa6cc,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nk97n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-868647ff47-9hc4f_openstack-operators(b9be4a5f-fbec-4e52-bb00-cbc08576e1fd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:32:03 crc kubenswrapper[4735]: E0215 20:32:03.125330 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" podUID="b9be4a5f-fbec-4e52-bb00-cbc08576e1fd" Feb 15 20:32:03 crc kubenswrapper[4735]: E0215 20:32:03.524211 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:90ad8fd8c1889b6be77925016532218eb6149d2c1c8535a5f9f1775c776fa6cc\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" podUID="b9be4a5f-fbec-4e52-bb00-cbc08576e1fd" Feb 15 20:32:04 crc kubenswrapper[4735]: E0215 20:32:04.304051 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867" Feb 15 20:32:04 crc kubenswrapper[4735]: E0215 20:32:04.304229 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4m9m4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-554564d7fc-rbhc6_openstack-operators(2b6ae856-4502-44cd-ad23-21577792b6a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:32:04 crc kubenswrapper[4735]: E0215 20:32:04.305355 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" podUID="2b6ae856-4502-44cd-ad23-21577792b6a0" Feb 15 20:32:04 crc kubenswrapper[4735]: E0215 20:32:04.532576 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" podUID="2b6ae856-4502-44cd-ad23-21577792b6a0" Feb 15 20:32:05 crc kubenswrapper[4735]: E0215 20:32:05.805287 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:2b8ab3063af4aaeed0198197aae6f391c6647ac686c94c85668537f1d5933979" Feb 15 20:32:05 crc kubenswrapper[4735]: E0215 20:32:05.806386 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:2b8ab3063af4aaeed0198197aae6f391c6647ac686c94c85668537f1d5933979,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mt64x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-5d946d989d-v6nvk_openstack-operators(98e57a69-0860-468e-898c-9ec42836d3d4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:32:05 crc kubenswrapper[4735]: E0215 20:32:05.807578 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" podUID="98e57a69-0860-468e-898c-9ec42836d3d4" Feb 15 20:32:06 crc kubenswrapper[4735]: E0215 20:32:06.485575 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838" Feb 15 20:32:06 crc kubenswrapper[4735]: E0215 20:32:06.486239 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fb4df,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-567668f5cf-t9nc4_openstack-operators(d8c734e9-fd8c-459d-982f-53b0a706d0a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:32:06 crc kubenswrapper[4735]: E0215 20:32:06.487524 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" podUID="d8c734e9-fd8c-459d-982f-53b0a706d0a0" Feb 15 20:32:06 crc kubenswrapper[4735]: E0215 20:32:06.546791 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838\\\"\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" podUID="d8c734e9-fd8c-459d-982f-53b0a706d0a0" Feb 15 20:32:06 crc kubenswrapper[4735]: E0215 20:32:06.548077 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:2b8ab3063af4aaeed0198197aae6f391c6647ac686c94c85668537f1d5933979\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" podUID="98e57a69-0860-468e-898c-9ec42836d3d4" Feb 15 20:32:08 crc kubenswrapper[4735]: E0215 20:32:08.545205 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1" Feb 15 20:32:08 crc kubenswrapper[4735]: E0215 20:32:08.545838 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r8f77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b4d948c87-ghs7m_openstack-operators(8802a6f8-206d-473d-854c-77e3c74389b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:32:08 crc kubenswrapper[4735]: E0215 20:32:08.547501 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" podUID="8802a6f8-206d-473d-854c-77e3c74389b5" Feb 15 20:32:08 crc kubenswrapper[4735]: E0215 20:32:08.558779 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" podUID="8802a6f8-206d-473d-854c-77e3c74389b5" Feb 15 20:32:10 crc kubenswrapper[4735]: I0215 20:32:10.379215 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:32:10 crc kubenswrapper[4735]: I0215 20:32:10.393603 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79556e2d-fb3e-4a41-a85c-38005174ba50-cert\") pod \"infra-operator-controller-manager-7676946c7c-5w649\" (UID: \"79556e2d-fb3e-4a41-a85c-38005174ba50\") " pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:32:10 crc kubenswrapper[4735]: I0215 20:32:10.579081 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:32:10 crc kubenswrapper[4735]: I0215 20:32:10.690622 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:32:10 crc kubenswrapper[4735]: I0215 20:32:10.695430 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9549fdaf-fe4c-4da2-92db-db91beb95450-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c48cdg5\" (UID: \"9549fdaf-fe4c-4da2-92db-db91beb95450\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:32:10 crc kubenswrapper[4735]: I0215 20:32:10.904502 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.204954 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.205354 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.214599 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-webhook-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.215759 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6fad7194-6e10-4dc9-ac3c-241722357e55-metrics-certs\") pod \"openstack-operator-controller-manager-79f9b74c8-4b426\" (UID: \"6fad7194-6e10-4dc9-ac3c-241722357e55\") " pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.353288 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7676946c7c-5w649"] Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.487284 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.630054 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" event={"ID":"14de9f25-fdbd-4d44-9f2c-440b256c4faf","Type":"ContainerStarted","Data":"3983143bad489fec8a693021ff9a287851d0701e4b6e9ed0f4a7650f05669d32"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.630383 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.646811 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" event={"ID":"79556e2d-fb3e-4a41-a85c-38005174ba50","Type":"ContainerStarted","Data":"a0b758ba754c634f160843b3b37fc678dff391163c1c06b141bb9b64220b981f"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.648888 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn68p" event={"ID":"147bc244-16d1-4e31-8ff4-f3b129803fd5","Type":"ContainerStarted","Data":"1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.657608 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" podStartSLOduration=7.014198673 podStartE2EDuration="33.657591656s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:41.11083405 +0000 UTC m=+908.976849673" lastFinishedPulling="2026-02-15 20:32:07.754227033 +0000 UTC m=+935.620242656" observedRunningTime="2026-02-15 20:32:11.656641441 +0000 UTC m=+939.522657064" watchObservedRunningTime="2026-02-15 20:32:11.657591656 +0000 UTC m=+939.523607279" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.669434 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" event={"ID":"c7d8cf5f-13c8-4857-96d2-0da2c88ed221","Type":"ContainerStarted","Data":"5fe86f1a34124b66ced7f4f46add8099b6783d332ade1b07fc0b8ffccffcaccd"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.670226 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.688694 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" event={"ID":"3fa6a777-e009-4e48-b05d-ac9b34acd21c","Type":"ContainerStarted","Data":"9fdbd331d6d3c7c09c5aeecc62eb9f1ba47e9ac76459770bb32e307ab0b223ea"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.689461 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.710720 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rn68p" podStartSLOduration=27.932672848 podStartE2EDuration="33.710707211s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:42.294736386 +0000 UTC m=+910.160752009" lastFinishedPulling="2026-02-15 20:31:48.072770749 +0000 UTC m=+915.938786372" observedRunningTime="2026-02-15 20:32:11.700986511 +0000 UTC m=+939.567002134" watchObservedRunningTime="2026-02-15 20:32:11.710707211 +0000 UTC m=+939.576722824" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.720524 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" event={"ID":"1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0","Type":"ContainerStarted","Data":"63e9cae723becea60d997cb773959a0b414f0757aa0e0e971df7663fdffca293"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.721199 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.748550 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" podStartSLOduration=6.202446403 podStartE2EDuration="33.748533467s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.980445101 +0000 UTC m=+908.846460724" lastFinishedPulling="2026-02-15 20:32:08.526532165 +0000 UTC m=+936.392547788" observedRunningTime="2026-02-15 20:32:11.747364525 +0000 UTC m=+939.613380148" watchObservedRunningTime="2026-02-15 20:32:11.748533467 +0000 UTC m=+939.614549090" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.764104 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" event={"ID":"26e7f3b0-6f34-491e-917a-e545d320ee89","Type":"ContainerStarted","Data":"c8865329beb3e2e655995eca8c2b70ef6328f9cf2f611056fd8595d7393436e7"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.764760 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.780671 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" event={"ID":"ab28d8d3-58f1-460c-a0ed-75f61ebc2c52","Type":"ContainerStarted","Data":"ed8357e6c957bdd31c891447d616fa56ff2a31b750a06ddf5e8f042a75392ad2"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.781366 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.812005 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" podStartSLOduration=6.122529507 podStartE2EDuration="33.811984858s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.838067641 +0000 UTC m=+908.704083264" lastFinishedPulling="2026-02-15 20:32:08.527523002 +0000 UTC m=+936.393538615" observedRunningTime="2026-02-15 20:32:11.809368318 +0000 UTC m=+939.675383941" watchObservedRunningTime="2026-02-15 20:32:11.811984858 +0000 UTC m=+939.678000471" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.832232 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" event={"ID":"99564ffe-1d8b-44b1-a625-4445d624a98f","Type":"ContainerStarted","Data":"8b9c4c7cfb4bc8a8fa2f59a677211618760afca4cad7b44e941f504d1782448b"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.832873 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.853192 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" event={"ID":"4fb2a283-35ca-4391-8e59-79a8fa7bb4f0","Type":"ContainerStarted","Data":"7431bf7e411afd2fadbf88c7e16fb38cb8a25cd0c4a1d39f2c75f9f32d560dae"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.853819 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.878150 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" event={"ID":"92aeecb8-b9fa-4db8-916e-f2dd800d2da3","Type":"ContainerStarted","Data":"fa570a33830734b896a4b03512f1540a36343730117e1a03dcad4066f3846062"} Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.878738 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" Feb 15 20:32:11 crc kubenswrapper[4735]: I0215 20:32:11.932137 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" podStartSLOduration=6.224877104 podStartE2EDuration="33.932117132s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.8201391 +0000 UTC m=+908.686154713" lastFinishedPulling="2026-02-15 20:32:08.527379118 +0000 UTC m=+936.393394741" observedRunningTime="2026-02-15 20:32:11.923332326 +0000 UTC m=+939.789347949" watchObservedRunningTime="2026-02-15 20:32:11.932117132 +0000 UTC m=+939.798132755" Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.245786 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" podStartSLOduration=6.840331187 podStartE2EDuration="34.245770058s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:41.122567904 +0000 UTC m=+908.988583527" lastFinishedPulling="2026-02-15 20:32:08.528006775 +0000 UTC m=+936.394022398" observedRunningTime="2026-02-15 20:32:12.076364353 +0000 UTC m=+939.942379976" watchObservedRunningTime="2026-02-15 20:32:12.245770058 +0000 UTC m=+940.111785671" Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.247977 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5"] Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.441181 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" podStartSLOduration=8.228408392 podStartE2EDuration="34.441163471s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.952569733 +0000 UTC m=+908.818585356" lastFinishedPulling="2026-02-15 20:32:07.165324812 +0000 UTC m=+935.031340435" observedRunningTime="2026-02-15 20:32:12.441101139 +0000 UTC m=+940.307116772" watchObservedRunningTime="2026-02-15 20:32:12.441163471 +0000 UTC m=+940.307179084" Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.444266 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" podStartSLOduration=7.676381751 podStartE2EDuration="34.444244523s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.987391538 +0000 UTC m=+908.853407161" lastFinishedPulling="2026-02-15 20:32:07.75525431 +0000 UTC m=+935.621269933" observedRunningTime="2026-02-15 20:32:12.304917165 +0000 UTC m=+940.170932778" watchObservedRunningTime="2026-02-15 20:32:12.444244523 +0000 UTC m=+940.310260146" Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.774231 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" podStartSLOduration=6.375132852 podStartE2EDuration="34.774216327s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.128822947 +0000 UTC m=+907.994838570" lastFinishedPulling="2026-02-15 20:32:08.527906422 +0000 UTC m=+936.393922045" observedRunningTime="2026-02-15 20:32:12.587127077 +0000 UTC m=+940.453142700" watchObservedRunningTime="2026-02-15 20:32:12.774216327 +0000 UTC m=+940.640231950" Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.780574 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" podStartSLOduration=6.436369548 podStartE2EDuration="34.780564717s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.18248484 +0000 UTC m=+908.048500463" lastFinishedPulling="2026-02-15 20:32:08.526680009 +0000 UTC m=+936.392695632" observedRunningTime="2026-02-15 20:32:12.773705934 +0000 UTC m=+940.639721557" watchObservedRunningTime="2026-02-15 20:32:12.780564717 +0000 UTC m=+940.646580340" Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.780996 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426"] Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.938062 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" event={"ID":"1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea","Type":"ContainerStarted","Data":"3cb8b9debd83a5e5c3c2aafd9d537b85d1df807ec3cd46eef8d786d99c1bcef0"} Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.938650 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.945924 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" event={"ID":"9549fdaf-fe4c-4da2-92db-db91beb95450","Type":"ContainerStarted","Data":"7bc3192ec20b2ae2df01253cc2234da3de90d7938b1d4373ce8157864bcdceda"} Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.947922 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" event={"ID":"e57b4eea-5c8f-4caa-9644-0994108974da","Type":"ContainerStarted","Data":"40a4f14ee4a9c7feaaa8d0f9892a42ff520a4f71a277c6cc587209ffea120cef"} Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.949338 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" event={"ID":"79e82794-8193-4eba-af98-f6e0ae43274b","Type":"ContainerStarted","Data":"be0f8360c9d20ecc5c6d7c196b1356db7bd6fa57bd86ea799b0f62af4ea8e632"} Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.949743 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.951833 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" event={"ID":"6fad7194-6e10-4dc9-ac3c-241722357e55","Type":"ContainerStarted","Data":"301b1d1bd269ed84efc40e1e90b47856e74db9caeb059d9c77619899526a7a9c"} Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.953147 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" event={"ID":"d8bf1d9d-df22-4b17-91b7-833052b5a337","Type":"ContainerStarted","Data":"52a63443d6d64738a7c86202acf46ca1c61b9924369f648433fa8db3123c4798"} Feb 15 20:32:12 crc kubenswrapper[4735]: I0215 20:32:12.953535 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" Feb 15 20:32:13 crc kubenswrapper[4735]: I0215 20:32:13.114346 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" podStartSLOduration=5.430626132 podStartE2EDuration="35.114330872s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:41.114556559 +0000 UTC m=+908.980572182" lastFinishedPulling="2026-02-15 20:32:10.798261289 +0000 UTC m=+938.664276922" observedRunningTime="2026-02-15 20:32:13.109956046 +0000 UTC m=+940.975971669" watchObservedRunningTime="2026-02-15 20:32:13.114330872 +0000 UTC m=+940.980346495" Feb 15 20:32:13 crc kubenswrapper[4735]: I0215 20:32:13.160558 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" podStartSLOduration=5.5026598159999995 podStartE2EDuration="35.160542233s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:41.15112122 +0000 UTC m=+909.017136843" lastFinishedPulling="2026-02-15 20:32:10.809003627 +0000 UTC m=+938.675019260" observedRunningTime="2026-02-15 20:32:13.159821003 +0000 UTC m=+941.025836626" watchObservedRunningTime="2026-02-15 20:32:13.160542233 +0000 UTC m=+941.026557856" Feb 15 20:32:13 crc kubenswrapper[4735]: I0215 20:32:13.207305 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" podStartSLOduration=5.511898173 podStartE2EDuration="35.207285146s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:41.151328276 +0000 UTC m=+909.017343899" lastFinishedPulling="2026-02-15 20:32:10.846715229 +0000 UTC m=+938.712730872" observedRunningTime="2026-02-15 20:32:13.187706191 +0000 UTC m=+941.053721814" watchObservedRunningTime="2026-02-15 20:32:13.207285146 +0000 UTC m=+941.073300759" Feb 15 20:32:13 crc kubenswrapper[4735]: I0215 20:32:13.227436 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9z8vz" podStartSLOduration=4.419757571 podStartE2EDuration="34.227421497s" podCreationTimestamp="2026-02-15 20:31:39 +0000 UTC" firstStartedPulling="2026-02-15 20:31:41.164603122 +0000 UTC m=+909.030618745" lastFinishedPulling="2026-02-15 20:32:10.972267048 +0000 UTC m=+938.838282671" observedRunningTime="2026-02-15 20:32:13.224034496 +0000 UTC m=+941.090050119" watchObservedRunningTime="2026-02-15 20:32:13.227421497 +0000 UTC m=+941.093437120" Feb 15 20:32:13 crc kubenswrapper[4735]: I0215 20:32:13.961128 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" event={"ID":"6fad7194-6e10-4dc9-ac3c-241722357e55","Type":"ContainerStarted","Data":"1a4228b00ea5c171f7df5ddacc5ef3794765be0e67ef310abac0960fe7d1e173"} Feb 15 20:32:14 crc kubenswrapper[4735]: I0215 20:32:14.968429 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:32:14 crc kubenswrapper[4735]: I0215 20:32:14.999357 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" podStartSLOduration=36.999341 podStartE2EDuration="36.999341s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:32:14.995563379 +0000 UTC m=+942.861579002" watchObservedRunningTime="2026-02-15 20:32:14.999341 +0000 UTC m=+942.865356623" Feb 15 20:32:17 crc kubenswrapper[4735]: I0215 20:32:17.986838 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" event={"ID":"2b6ae856-4502-44cd-ad23-21577792b6a0","Type":"ContainerStarted","Data":"d3b64ec98445df7e7406152e056ded4ea0e55f412a0f98ef339d93c8ba1fc5e1"} Feb 15 20:32:17 crc kubenswrapper[4735]: I0215 20:32:17.988251 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:17.989601 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" event={"ID":"31865132-118d-4f4c-bf0b-44ab13060882","Type":"ContainerStarted","Data":"058b6480fa618ce4b6852ee6e1f0952b5fa51b5554f6b073f2cdc9bca3e75433"} Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:17.990185 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:17.991414 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" event={"ID":"1094ce7c-b104-40d4-8e0f-3ee695bfd6cd","Type":"ContainerStarted","Data":"f324a63aee1b9803f53dde2d6595f88e0eea0c2b3aa053546ecbeeedd8f7d7c3"} Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:17.991760 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.001229 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" event={"ID":"79556e2d-fb3e-4a41-a85c-38005174ba50","Type":"ContainerStarted","Data":"a1d582c3dff5ce19b47c5ad841c35e9711b9819763573ee003208efe21081408"} Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.001401 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.014159 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" event={"ID":"9549fdaf-fe4c-4da2-92db-db91beb95450","Type":"ContainerStarted","Data":"44fd3f3d61bd3f6856116ed4e545f623d1bc6edc9f618b8d41757a77d008a996"} Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.014754 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.022720 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" event={"ID":"b9be4a5f-fbec-4e52-bb00-cbc08576e1fd","Type":"ContainerStarted","Data":"6d015a5a5ee31c76bf8fa9b6f6fd048e1a09d248b6c1de3f34b629beaf7a951a"} Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.023281 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.037797 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" podStartSLOduration=3.368808831 podStartE2EDuration="40.037779696s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.76685298 +0000 UTC m=+908.632868603" lastFinishedPulling="2026-02-15 20:32:17.435823825 +0000 UTC m=+945.301839468" observedRunningTime="2026-02-15 20:32:18.030243214 +0000 UTC m=+945.896258837" watchObservedRunningTime="2026-02-15 20:32:18.037779696 +0000 UTC m=+945.903795309" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.063224 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" podStartSLOduration=2.54376348 podStartE2EDuration="40.063200948s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:39.917821045 +0000 UTC m=+907.783836668" lastFinishedPulling="2026-02-15 20:32:17.437258513 +0000 UTC m=+945.303274136" observedRunningTime="2026-02-15 20:32:18.061011059 +0000 UTC m=+945.927026682" watchObservedRunningTime="2026-02-15 20:32:18.063200948 +0000 UTC m=+945.929216571" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.094878 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" podStartSLOduration=2.621301381 podStartE2EDuration="40.094862598s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:39.961170628 +0000 UTC m=+907.827186251" lastFinishedPulling="2026-02-15 20:32:17.434731835 +0000 UTC m=+945.300747468" observedRunningTime="2026-02-15 20:32:18.087827219 +0000 UTC m=+945.953842832" watchObservedRunningTime="2026-02-15 20:32:18.094862598 +0000 UTC m=+945.960878221" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.127437 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" podStartSLOduration=34.94384655 podStartE2EDuration="40.127419672s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:32:12.250258939 +0000 UTC m=+940.116274562" lastFinishedPulling="2026-02-15 20:32:17.433832061 +0000 UTC m=+945.299847684" observedRunningTime="2026-02-15 20:32:18.124898804 +0000 UTC m=+945.990914417" watchObservedRunningTime="2026-02-15 20:32:18.127419672 +0000 UTC m=+945.993435295" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.157521 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" podStartSLOduration=34.166402487 podStartE2EDuration="40.157506538s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:32:11.413580549 +0000 UTC m=+939.279596172" lastFinishedPulling="2026-02-15 20:32:17.4046846 +0000 UTC m=+945.270700223" observedRunningTime="2026-02-15 20:32:18.155791473 +0000 UTC m=+946.021807096" watchObservedRunningTime="2026-02-15 20:32:18.157506538 +0000 UTC m=+946.023522161" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.179146 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" podStartSLOduration=6.183733181 podStartE2EDuration="40.179131709s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.761004023 +0000 UTC m=+908.627019646" lastFinishedPulling="2026-02-15 20:32:14.756402551 +0000 UTC m=+942.622418174" observedRunningTime="2026-02-15 20:32:18.174882405 +0000 UTC m=+946.040898028" watchObservedRunningTime="2026-02-15 20:32:18.179131709 +0000 UTC m=+946.045147332" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.555772 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-vddvx" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.566033 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987464f4-gvnkh" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.634557 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-v4wpf" Feb 15 20:32:18 crc kubenswrapper[4735]: I0215 20:32:18.985088 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-jwrfj" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.020909 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-mh89r" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.029058 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" event={"ID":"d8c734e9-fd8c-459d-982f-53b0a706d0a0","Type":"ContainerStarted","Data":"1e7b5aa2c72f7e8f2e2e30fd930e2d561ef8479859e1085f3939f64692d5e0a1"} Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.029638 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.031047 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" event={"ID":"98e57a69-0860-468e-898c-9ec42836d3d4","Type":"ContainerStarted","Data":"fe40bb15d513302a56b20e5ced64c58d0f3155f531241549f27cdabc7d48f485"} Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.031351 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.059198 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-w8vz9" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.081793 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-f8s72" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.132655 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" podStartSLOduration=4.007184389 podStartE2EDuration="41.132638402s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.831217607 +0000 UTC m=+908.697233230" lastFinishedPulling="2026-02-15 20:32:17.95667162 +0000 UTC m=+945.822687243" observedRunningTime="2026-02-15 20:32:19.130398343 +0000 UTC m=+946.996413966" watchObservedRunningTime="2026-02-15 20:32:19.132638402 +0000 UTC m=+946.998654025" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.163834 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" podStartSLOduration=3.067970971 podStartE2EDuration="41.16381718s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.265273633 +0000 UTC m=+908.131289256" lastFinishedPulling="2026-02-15 20:32:18.361119842 +0000 UTC m=+946.227135465" observedRunningTime="2026-02-15 20:32:19.161643281 +0000 UTC m=+947.027658904" watchObservedRunningTime="2026-02-15 20:32:19.16381718 +0000 UTC m=+947.029832803" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.254890 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9trcb" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.389478 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-vbr6n" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.398464 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-lfj2m" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.429292 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7866795846-r6bz7" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.562562 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-d65l6" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.650439 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.650742 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.679775 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:32:19 crc kubenswrapper[4735]: I0215 20:32:19.679832 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:32:20 crc kubenswrapper[4735]: I0215 20:32:20.699186 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rn68p" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerName="registry-server" probeResult="failure" output=< Feb 15 20:32:20 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:32:20 crc kubenswrapper[4735]: > Feb 15 20:32:21 crc kubenswrapper[4735]: I0215 20:32:21.494754 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-79f9b74c8-4b426" Feb 15 20:32:22 crc kubenswrapper[4735]: I0215 20:32:22.051638 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" event={"ID":"8802a6f8-206d-473d-854c-77e3c74389b5","Type":"ContainerStarted","Data":"49e5d3bdc3a152731168876da38d3b1586120f4359df2783d0186892c1a1a9f4"} Feb 15 20:32:22 crc kubenswrapper[4735]: I0215 20:32:22.051853 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" Feb 15 20:32:22 crc kubenswrapper[4735]: I0215 20:32:22.072891 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" podStartSLOduration=3.462968478 podStartE2EDuration="44.072874144s" podCreationTimestamp="2026-02-15 20:31:38 +0000 UTC" firstStartedPulling="2026-02-15 20:31:40.76126043 +0000 UTC m=+908.627276053" lastFinishedPulling="2026-02-15 20:32:21.371166096 +0000 UTC m=+949.237181719" observedRunningTime="2026-02-15 20:32:22.068567109 +0000 UTC m=+949.934582732" watchObservedRunningTime="2026-02-15 20:32:22.072874144 +0000 UTC m=+949.938889767" Feb 15 20:32:28 crc kubenswrapper[4735]: I0215 20:32:28.513355 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-9hc4f" Feb 15 20:32:28 crc kubenswrapper[4735]: I0215 20:32:28.530672 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-v6nvk" Feb 15 20:32:28 crc kubenswrapper[4735]: I0215 20:32:28.669385 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-fslrr" Feb 15 20:32:28 crc kubenswrapper[4735]: I0215 20:32:28.816094 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-rbhc6" Feb 15 20:32:28 crc kubenswrapper[4735]: I0215 20:32:28.976136 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ghs7m" Feb 15 20:32:28 crc kubenswrapper[4735]: I0215 20:32:28.985422 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-4hlsd" Feb 15 20:32:29 crc kubenswrapper[4735]: I0215 20:32:29.049933 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9nc4" Feb 15 20:32:29 crc kubenswrapper[4735]: I0215 20:32:29.706353 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:32:29 crc kubenswrapper[4735]: I0215 20:32:29.758239 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:32:29 crc kubenswrapper[4735]: I0215 20:32:29.948989 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rn68p"] Feb 15 20:32:30 crc kubenswrapper[4735]: I0215 20:32:30.585734 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7676946c7c-5w649" Feb 15 20:32:30 crc kubenswrapper[4735]: I0215 20:32:30.914156 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c48cdg5" Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.116184 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rn68p" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerName="registry-server" containerID="cri-o://1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc" gracePeriod=2 Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.522359 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.584602 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtvgs\" (UniqueName: \"kubernetes.io/projected/147bc244-16d1-4e31-8ff4-f3b129803fd5-kube-api-access-wtvgs\") pod \"147bc244-16d1-4e31-8ff4-f3b129803fd5\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.584904 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-utilities\") pod \"147bc244-16d1-4e31-8ff4-f3b129803fd5\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.585050 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-catalog-content\") pod \"147bc244-16d1-4e31-8ff4-f3b129803fd5\" (UID: \"147bc244-16d1-4e31-8ff4-f3b129803fd5\") " Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.585686 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-utilities" (OuterVolumeSpecName: "utilities") pod "147bc244-16d1-4e31-8ff4-f3b129803fd5" (UID: "147bc244-16d1-4e31-8ff4-f3b129803fd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.590389 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147bc244-16d1-4e31-8ff4-f3b129803fd5-kube-api-access-wtvgs" (OuterVolumeSpecName: "kube-api-access-wtvgs") pod "147bc244-16d1-4e31-8ff4-f3b129803fd5" (UID: "147bc244-16d1-4e31-8ff4-f3b129803fd5"). InnerVolumeSpecName "kube-api-access-wtvgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.645488 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "147bc244-16d1-4e31-8ff4-f3b129803fd5" (UID: "147bc244-16d1-4e31-8ff4-f3b129803fd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.686800 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.686843 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtvgs\" (UniqueName: \"kubernetes.io/projected/147bc244-16d1-4e31-8ff4-f3b129803fd5-kube-api-access-wtvgs\") on node \"crc\" DevicePath \"\"" Feb 15 20:32:31 crc kubenswrapper[4735]: I0215 20:32:31.686859 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147bc244-16d1-4e31-8ff4-f3b129803fd5-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.125331 4735 generic.go:334] "Generic (PLEG): container finished" podID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerID="1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc" exitCode=0 Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.125372 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn68p" event={"ID":"147bc244-16d1-4e31-8ff4-f3b129803fd5","Type":"ContainerDied","Data":"1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc"} Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.125396 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn68p" event={"ID":"147bc244-16d1-4e31-8ff4-f3b129803fd5","Type":"ContainerDied","Data":"b2fde80dba9d1ba124d35fe0293fea09655af2c0b0b31cd2249b637e5004479f"} Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.125416 4735 scope.go:117] "RemoveContainer" containerID="1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.125536 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn68p" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.155604 4735 scope.go:117] "RemoveContainer" containerID="5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.170539 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rn68p"] Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.176116 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rn68p"] Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.187911 4735 scope.go:117] "RemoveContainer" containerID="1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.205045 4735 scope.go:117] "RemoveContainer" containerID="1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc" Feb 15 20:32:32 crc kubenswrapper[4735]: E0215 20:32:32.205429 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc\": container with ID starting with 1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc not found: ID does not exist" containerID="1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.205477 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc"} err="failed to get container status \"1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc\": rpc error: code = NotFound desc = could not find container \"1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc\": container with ID starting with 1cd1258f11c2cdc6bc98c6bdcb7a220be46a10903b5dce3db90425d1da47edfc not found: ID does not exist" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.205496 4735 scope.go:117] "RemoveContainer" containerID="5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f" Feb 15 20:32:32 crc kubenswrapper[4735]: E0215 20:32:32.205781 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f\": container with ID starting with 5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f not found: ID does not exist" containerID="5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.205886 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f"} err="failed to get container status \"5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f\": rpc error: code = NotFound desc = could not find container \"5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f\": container with ID starting with 5f9b6291aca575ba2f2c3f9006604caa233a7c683eb1fbd663e40f3bec481d5f not found: ID does not exist" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.206000 4735 scope.go:117] "RemoveContainer" containerID="1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99" Feb 15 20:32:32 crc kubenswrapper[4735]: E0215 20:32:32.206359 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99\": container with ID starting with 1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99 not found: ID does not exist" containerID="1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.206409 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99"} err="failed to get container status \"1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99\": rpc error: code = NotFound desc = could not find container \"1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99\": container with ID starting with 1fb7f6d7e5eb0078a2264db6a9587a4918dbe53dfa29013cf06909083fb4bc99 not found: ID does not exist" Feb 15 20:32:32 crc kubenswrapper[4735]: I0215 20:32:32.933865 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" path="/var/lib/kubelet/pods/147bc244-16d1-4e31-8ff4-f3b129803fd5/volumes" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.737037 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lz8md"] Feb 15 20:32:46 crc kubenswrapper[4735]: E0215 20:32:46.737830 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerName="extract-content" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.737844 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerName="extract-content" Feb 15 20:32:46 crc kubenswrapper[4735]: E0215 20:32:46.737867 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerName="registry-server" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.737874 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerName="registry-server" Feb 15 20:32:46 crc kubenswrapper[4735]: E0215 20:32:46.737908 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerName="extract-utilities" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.737917 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerName="extract-utilities" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.738065 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="147bc244-16d1-4e31-8ff4-f3b129803fd5" containerName="registry-server" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.738735 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.754956 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.755167 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-jzbt9" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.755361 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.755848 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.763543 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lz8md"] Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.824424 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qhds\" (UniqueName: \"kubernetes.io/projected/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-kube-api-access-5qhds\") pod \"dnsmasq-dns-675f4bcbfc-lz8md\" (UID: \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.824521 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-config\") pod \"dnsmasq-dns-675f4bcbfc-lz8md\" (UID: \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.869316 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d4n6w"] Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.870604 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.875266 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.895368 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d4n6w"] Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.926433 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-config\") pod \"dnsmasq-dns-78dd6ddcc-d4n6w\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.926493 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-config\") pod \"dnsmasq-dns-675f4bcbfc-lz8md\" (UID: \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.926545 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qhds\" (UniqueName: \"kubernetes.io/projected/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-kube-api-access-5qhds\") pod \"dnsmasq-dns-675f4bcbfc-lz8md\" (UID: \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.926577 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8fvn\" (UniqueName: \"kubernetes.io/projected/49cf9284-c3f4-44a6-8625-2fecc515065a-kube-api-access-b8fvn\") pod \"dnsmasq-dns-78dd6ddcc-d4n6w\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.926621 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-d4n6w\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.927658 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-config\") pod \"dnsmasq-dns-675f4bcbfc-lz8md\" (UID: \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:32:46 crc kubenswrapper[4735]: I0215 20:32:46.951593 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qhds\" (UniqueName: \"kubernetes.io/projected/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-kube-api-access-5qhds\") pod \"dnsmasq-dns-675f4bcbfc-lz8md\" (UID: \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.028181 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8fvn\" (UniqueName: \"kubernetes.io/projected/49cf9284-c3f4-44a6-8625-2fecc515065a-kube-api-access-b8fvn\") pod \"dnsmasq-dns-78dd6ddcc-d4n6w\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.028241 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-d4n6w\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.028291 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-config\") pod \"dnsmasq-dns-78dd6ddcc-d4n6w\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.029292 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-d4n6w\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.029752 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-config\") pod \"dnsmasq-dns-78dd6ddcc-d4n6w\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.047541 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8fvn\" (UniqueName: \"kubernetes.io/projected/49cf9284-c3f4-44a6-8625-2fecc515065a-kube-api-access-b8fvn\") pod \"dnsmasq-dns-78dd6ddcc-d4n6w\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.070817 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.187171 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.496392 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lz8md"] Feb 15 20:32:47 crc kubenswrapper[4735]: W0215 20:32:47.500690 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod587df3d0_fd47_4cc8_8d59_bd9a34c19a79.slice/crio-60613e892c1d79212af45d0e5f7840a591f3cbcf28c02f186430bdcd26a136d1 WatchSource:0}: Error finding container 60613e892c1d79212af45d0e5f7840a591f3cbcf28c02f186430bdcd26a136d1: Status 404 returned error can't find the container with id 60613e892c1d79212af45d0e5f7840a591f3cbcf28c02f186430bdcd26a136d1 Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.503046 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 20:32:47 crc kubenswrapper[4735]: I0215 20:32:47.609261 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d4n6w"] Feb 15 20:32:47 crc kubenswrapper[4735]: W0215 20:32:47.611182 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49cf9284_c3f4_44a6_8625_2fecc515065a.slice/crio-9811f7ac0749b95d4332649c486e6d30baada140669b9cdc665c457f52e3cbe5 WatchSource:0}: Error finding container 9811f7ac0749b95d4332649c486e6d30baada140669b9cdc665c457f52e3cbe5: Status 404 returned error can't find the container with id 9811f7ac0749b95d4332649c486e6d30baada140669b9cdc665c457f52e3cbe5 Feb 15 20:32:48 crc kubenswrapper[4735]: I0215 20:32:48.247661 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" event={"ID":"587df3d0-fd47-4cc8-8d59-bd9a34c19a79","Type":"ContainerStarted","Data":"60613e892c1d79212af45d0e5f7840a591f3cbcf28c02f186430bdcd26a136d1"} Feb 15 20:32:48 crc kubenswrapper[4735]: I0215 20:32:48.249105 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" event={"ID":"49cf9284-c3f4-44a6-8625-2fecc515065a","Type":"ContainerStarted","Data":"9811f7ac0749b95d4332649c486e6d30baada140669b9cdc665c457f52e3cbe5"} Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.495446 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lz8md"] Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.521890 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-drgd2"] Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.522909 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.570648 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-drgd2\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.570738 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c2ld\" (UniqueName: \"kubernetes.io/projected/1dfdacbc-8183-4181-854b-bb0be9230eb8-kube-api-access-6c2ld\") pod \"dnsmasq-dns-666b6646f7-drgd2\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.570774 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-config\") pod \"dnsmasq-dns-666b6646f7-drgd2\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.604437 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-drgd2"] Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.672130 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-drgd2\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.672230 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c2ld\" (UniqueName: \"kubernetes.io/projected/1dfdacbc-8183-4181-854b-bb0be9230eb8-kube-api-access-6c2ld\") pod \"dnsmasq-dns-666b6646f7-drgd2\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.672268 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-config\") pod \"dnsmasq-dns-666b6646f7-drgd2\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.673289 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-config\") pod \"dnsmasq-dns-666b6646f7-drgd2\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.673991 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-drgd2\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.681316 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.681402 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.681469 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.684195 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"776a88a7dd582588d5537251275510c6e25c50e714d7281e0f68c913b0f4fe42"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.684271 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://776a88a7dd582588d5537251275510c6e25c50e714d7281e0f68c913b0f4fe42" gracePeriod=600 Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.724688 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c2ld\" (UniqueName: \"kubernetes.io/projected/1dfdacbc-8183-4181-854b-bb0be9230eb8-kube-api-access-6c2ld\") pod \"dnsmasq-dns-666b6646f7-drgd2\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.860903 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d4n6w"] Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.869241 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.877789 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jzftf"] Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.882537 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.901642 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jzftf"] Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.978223 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mlbv\" (UniqueName: \"kubernetes.io/projected/d5382c75-62d2-412c-ab22-e1a8a83a2e71-kube-api-access-4mlbv\") pod \"dnsmasq-dns-57d769cc4f-jzftf\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.978337 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-config\") pod \"dnsmasq-dns-57d769cc4f-jzftf\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:49 crc kubenswrapper[4735]: I0215 20:32:49.978358 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-jzftf\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.083646 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-config\") pod \"dnsmasq-dns-57d769cc4f-jzftf\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.083691 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-jzftf\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.083796 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mlbv\" (UniqueName: \"kubernetes.io/projected/d5382c75-62d2-412c-ab22-e1a8a83a2e71-kube-api-access-4mlbv\") pod \"dnsmasq-dns-57d769cc4f-jzftf\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.085011 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-config\") pod \"dnsmasq-dns-57d769cc4f-jzftf\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.085060 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-jzftf\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.115248 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mlbv\" (UniqueName: \"kubernetes.io/projected/d5382c75-62d2-412c-ab22-e1a8a83a2e71-kube-api-access-4mlbv\") pod \"dnsmasq-dns-57d769cc4f-jzftf\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.208561 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.286889 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="776a88a7dd582588d5537251275510c6e25c50e714d7281e0f68c913b0f4fe42" exitCode=0 Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.287211 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"776a88a7dd582588d5537251275510c6e25c50e714d7281e0f68c913b0f4fe42"} Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.287236 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"94228183c3d1abcbd5b788cf2f92ec757e11f8d18a65dfdd44912dcfc7991e88"} Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.287252 4735 scope.go:117] "RemoveContainer" containerID="8137201feee74b50c12e4ac603bd7a608ffb1bf6f8ae2e8df072ae67e560dfe2" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.687729 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.689332 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.693966 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.694530 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-g2d24" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.697388 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.697529 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.697686 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.697791 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.699032 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.720883 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.768687 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-drgd2"] Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802480 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802522 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802560 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc3c285a-910a-4cb1-968e-636216557d09-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802582 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802606 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802620 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52cqz\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-kube-api-access-52cqz\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802643 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc3c285a-910a-4cb1-968e-636216557d09-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802711 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-config-data\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802729 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802744 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.802777 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.845281 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jzftf"] Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.904380 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.904513 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.904535 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52cqz\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-kube-api-access-52cqz\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.904560 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc3c285a-910a-4cb1-968e-636216557d09-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.904597 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-config-data\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.904613 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.904629 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.904654 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.904797 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.906543 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-config-data\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.906568 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.906738 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.906775 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.906815 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc3c285a-910a-4cb1-968e-636216557d09-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.907170 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.908016 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.912775 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.919367 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc3c285a-910a-4cb1-968e-636216557d09-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.919530 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.921266 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc3c285a-910a-4cb1-968e-636216557d09-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.930114 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52cqz\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-kube-api-access-52cqz\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.944792 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:50 crc kubenswrapper[4735]: I0215 20:32:50.951274 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " pod="openstack/rabbitmq-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.029430 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.030528 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.039789 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.040066 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.040228 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-dvwjn" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.040740 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.040875 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.040976 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.041077 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.058549 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.070545 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.110853 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.110892 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.110921 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.110964 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c7b91a3-10d6-465d-aff4-4855d697b54f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.110992 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.111008 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9727r\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-kube-api-access-9727r\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.118958 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.119001 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.119111 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.119170 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.119227 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c7b91a3-10d6-465d-aff4-4855d697b54f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221354 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221454 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221543 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c7b91a3-10d6-465d-aff4-4855d697b54f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221626 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221648 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221707 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221760 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c7b91a3-10d6-465d-aff4-4855d697b54f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221810 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.221851 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.222976 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9727r\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-kube-api-access-9727r\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.223002 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.222467 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.222136 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.223229 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.222451 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.223646 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.228059 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.230679 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c7b91a3-10d6-465d-aff4-4855d697b54f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.232437 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.237726 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c7b91a3-10d6-465d-aff4-4855d697b54f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.243527 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9727r\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-kube-api-access-9727r\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.256594 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.307436 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-drgd2" event={"ID":"1dfdacbc-8183-4181-854b-bb0be9230eb8","Type":"ContainerStarted","Data":"78d7757a973808b804921d881e142efa128ff6c1403eea17a7c63b53ec5864d7"} Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.313087 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" event={"ID":"d5382c75-62d2-412c-ab22-e1a8a83a2e71","Type":"ContainerStarted","Data":"1d4bb0fec04c39257007a31c1763cdba4e62f479b113e32440eb8b61d9f4eff2"} Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.358849 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.617954 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 15 20:32:51 crc kubenswrapper[4735]: I0215 20:32:51.909446 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 15 20:32:51 crc kubenswrapper[4735]: W0215 20:32:51.919432 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c7b91a3_10d6_465d_aff4_4855d697b54f.slice/crio-6d058253789e53d5e75363ee2b4962a083c87a62523c4e2669f5d420caf190a5 WatchSource:0}: Error finding container 6d058253789e53d5e75363ee2b4962a083c87a62523c4e2669f5d420caf190a5: Status 404 returned error can't find the container with id 6d058253789e53d5e75363ee2b4962a083c87a62523c4e2669f5d420caf190a5 Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.189609 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.197660 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.205164 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.205578 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.208271 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.208548 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.208802 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-rstg5" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.216740 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.244138 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f1ba-392a-4523-927c-405fab5bd619-operator-scripts\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.244634 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.244727 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aa84f1ba-392a-4523-927c-405fab5bd619-kolla-config\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.244786 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/aa84f1ba-392a-4523-927c-405fab5bd619-config-data-generated\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.244864 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa84f1ba-392a-4523-927c-405fab5bd619-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.244925 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa84f1ba-392a-4523-927c-405fab5bd619-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.244986 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svvvr\" (UniqueName: \"kubernetes.io/projected/aa84f1ba-392a-4523-927c-405fab5bd619-kube-api-access-svvvr\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.245008 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/aa84f1ba-392a-4523-927c-405fab5bd619-config-data-default\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.322055 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc3c285a-910a-4cb1-968e-636216557d09","Type":"ContainerStarted","Data":"6aed3fa6dea6824209d54628ebf241ba43a0520b0dfd95cfc059dd181621c81c"} Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.323394 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5c7b91a3-10d6-465d-aff4-4855d697b54f","Type":"ContainerStarted","Data":"6d058253789e53d5e75363ee2b4962a083c87a62523c4e2669f5d420caf190a5"} Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.346602 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f1ba-392a-4523-927c-405fab5bd619-operator-scripts\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.346645 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.346707 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aa84f1ba-392a-4523-927c-405fab5bd619-kolla-config\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.346729 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/aa84f1ba-392a-4523-927c-405fab5bd619-config-data-generated\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.346760 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa84f1ba-392a-4523-927c-405fab5bd619-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.346793 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa84f1ba-392a-4523-927c-405fab5bd619-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.346819 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svvvr\" (UniqueName: \"kubernetes.io/projected/aa84f1ba-392a-4523-927c-405fab5bd619-kube-api-access-svvvr\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.346841 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/aa84f1ba-392a-4523-927c-405fab5bd619-config-data-default\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.347847 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/aa84f1ba-392a-4523-927c-405fab5bd619-config-data-default\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.348678 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/aa84f1ba-392a-4523-927c-405fab5bd619-config-data-generated\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.349006 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.349168 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aa84f1ba-392a-4523-927c-405fab5bd619-kolla-config\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.349750 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f1ba-392a-4523-927c-405fab5bd619-operator-scripts\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.368845 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa84f1ba-392a-4523-927c-405fab5bd619-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.370514 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa84f1ba-392a-4523-927c-405fab5bd619-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.388156 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svvvr\" (UniqueName: \"kubernetes.io/projected/aa84f1ba-392a-4523-927c-405fab5bd619-kube-api-access-svvvr\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.396447 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"aa84f1ba-392a-4523-927c-405fab5bd619\") " pod="openstack/openstack-galera-0" Feb 15 20:32:52 crc kubenswrapper[4735]: I0215 20:32:52.542323 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.371460 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 15 20:32:53 crc kubenswrapper[4735]: W0215 20:32:53.382277 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa84f1ba_392a_4523_927c_405fab5bd619.slice/crio-e6a0fad58e50890649942525e50a386b6a67a4089586ff2125ec1303b4aa3690 WatchSource:0}: Error finding container e6a0fad58e50890649942525e50a386b6a67a4089586ff2125ec1303b4aa3690: Status 404 returned error can't find the container with id e6a0fad58e50890649942525e50a386b6a67a4089586ff2125ec1303b4aa3690 Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.501363 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.503693 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.516821 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-mzdkb" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.517323 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.517328 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.517700 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.524260 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.674437 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f87d118-86d0-477e-a735-d2cd281b334a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.674776 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.674806 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1f87d118-86d0-477e-a735-d2cd281b334a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.674857 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f87d118-86d0-477e-a735-d2cd281b334a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.674890 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbxfc\" (UniqueName: \"kubernetes.io/projected/1f87d118-86d0-477e-a735-d2cd281b334a-kube-api-access-zbxfc\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.674931 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1f87d118-86d0-477e-a735-d2cd281b334a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.674985 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1f87d118-86d0-477e-a735-d2cd281b334a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.675005 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f87d118-86d0-477e-a735-d2cd281b334a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.776855 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f87d118-86d0-477e-a735-d2cd281b334a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.776901 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f87d118-86d0-477e-a735-d2cd281b334a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.777094 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.777220 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1f87d118-86d0-477e-a735-d2cd281b334a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.777344 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f87d118-86d0-477e-a735-d2cd281b334a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.777392 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbxfc\" (UniqueName: \"kubernetes.io/projected/1f87d118-86d0-477e-a735-d2cd281b334a-kube-api-access-zbxfc\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.777553 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1f87d118-86d0-477e-a735-d2cd281b334a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.777693 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1f87d118-86d0-477e-a735-d2cd281b334a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.777781 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.779120 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1f87d118-86d0-477e-a735-d2cd281b334a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.779403 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1f87d118-86d0-477e-a735-d2cd281b334a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.779419 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1f87d118-86d0-477e-a735-d2cd281b334a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.781973 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f87d118-86d0-477e-a735-d2cd281b334a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.794237 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f87d118-86d0-477e-a735-d2cd281b334a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.799876 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbxfc\" (UniqueName: \"kubernetes.io/projected/1f87d118-86d0-477e-a735-d2cd281b334a-kube-api-access-zbxfc\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.811667 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f87d118-86d0-477e-a735-d2cd281b334a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.816141 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1f87d118-86d0-477e-a735-d2cd281b334a\") " pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.840380 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.864696 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.868795 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.871311 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.871557 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-r5p94" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.871745 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.907792 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.983857 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a34b649-4566-4a07-846e-e97886eec195-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.983912 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a34b649-4566-4a07-846e-e97886eec195-config-data\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.984158 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a34b649-4566-4a07-846e-e97886eec195-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.984204 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bp78\" (UniqueName: \"kubernetes.io/projected/6a34b649-4566-4a07-846e-e97886eec195-kube-api-access-8bp78\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:53 crc kubenswrapper[4735]: I0215 20:32:53.984375 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6a34b649-4566-4a07-846e-e97886eec195-kolla-config\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.085754 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a34b649-4566-4a07-846e-e97886eec195-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.085795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bp78\" (UniqueName: \"kubernetes.io/projected/6a34b649-4566-4a07-846e-e97886eec195-kube-api-access-8bp78\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.085838 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6a34b649-4566-4a07-846e-e97886eec195-kolla-config\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.085865 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a34b649-4566-4a07-846e-e97886eec195-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.085890 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a34b649-4566-4a07-846e-e97886eec195-config-data\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.086588 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a34b649-4566-4a07-846e-e97886eec195-config-data\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.088292 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6a34b649-4566-4a07-846e-e97886eec195-kolla-config\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.091315 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a34b649-4566-4a07-846e-e97886eec195-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.096817 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a34b649-4566-4a07-846e-e97886eec195-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.101472 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bp78\" (UniqueName: \"kubernetes.io/projected/6a34b649-4566-4a07-846e-e97886eec195-kube-api-access-8bp78\") pod \"memcached-0\" (UID: \"6a34b649-4566-4a07-846e-e97886eec195\") " pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.228337 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.348884 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"aa84f1ba-392a-4523-927c-405fab5bd619","Type":"ContainerStarted","Data":"e6a0fad58e50890649942525e50a386b6a67a4089586ff2125ec1303b4aa3690"} Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.506855 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 15 20:32:54 crc kubenswrapper[4735]: I0215 20:32:54.832234 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 15 20:32:54 crc kubenswrapper[4735]: W0215 20:32:54.845307 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a34b649_4566_4a07_846e_e97886eec195.slice/crio-d42e5abf437600cccbe41da6574270682232cb1cd127b578c8dcfe4c95247a5d WatchSource:0}: Error finding container d42e5abf437600cccbe41da6574270682232cb1cd127b578c8dcfe4c95247a5d: Status 404 returned error can't find the container with id d42e5abf437600cccbe41da6574270682232cb1cd127b578c8dcfe4c95247a5d Feb 15 20:32:55 crc kubenswrapper[4735]: I0215 20:32:55.389094 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6a34b649-4566-4a07-846e-e97886eec195","Type":"ContainerStarted","Data":"d42e5abf437600cccbe41da6574270682232cb1cd127b578c8dcfe4c95247a5d"} Feb 15 20:32:55 crc kubenswrapper[4735]: I0215 20:32:55.397502 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1f87d118-86d0-477e-a735-d2cd281b334a","Type":"ContainerStarted","Data":"8b2c830a989931c449d1cbe5371930d8c08d3ec8a3b48f55337f45b80f00759f"} Feb 15 20:32:56 crc kubenswrapper[4735]: I0215 20:32:56.474631 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 15 20:32:56 crc kubenswrapper[4735]: I0215 20:32:56.475682 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 15 20:32:56 crc kubenswrapper[4735]: I0215 20:32:56.492418 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-l5pw6" Feb 15 20:32:56 crc kubenswrapper[4735]: I0215 20:32:56.514180 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 15 20:32:56 crc kubenswrapper[4735]: I0215 20:32:56.645628 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhwzq\" (UniqueName: \"kubernetes.io/projected/a01ba16b-aa0c-4811-8223-9f4c083fdeab-kube-api-access-hhwzq\") pod \"kube-state-metrics-0\" (UID: \"a01ba16b-aa0c-4811-8223-9f4c083fdeab\") " pod="openstack/kube-state-metrics-0" Feb 15 20:32:56 crc kubenswrapper[4735]: I0215 20:32:56.746929 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhwzq\" (UniqueName: \"kubernetes.io/projected/a01ba16b-aa0c-4811-8223-9f4c083fdeab-kube-api-access-hhwzq\") pod \"kube-state-metrics-0\" (UID: \"a01ba16b-aa0c-4811-8223-9f4c083fdeab\") " pod="openstack/kube-state-metrics-0" Feb 15 20:32:56 crc kubenswrapper[4735]: I0215 20:32:56.769110 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhwzq\" (UniqueName: \"kubernetes.io/projected/a01ba16b-aa0c-4811-8223-9f4c083fdeab-kube-api-access-hhwzq\") pod \"kube-state-metrics-0\" (UID: \"a01ba16b-aa0c-4811-8223-9f4c083fdeab\") " pod="openstack/kube-state-metrics-0" Feb 15 20:32:56 crc kubenswrapper[4735]: I0215 20:32:56.801326 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 15 20:32:57 crc kubenswrapper[4735]: I0215 20:32:57.412323 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 15 20:32:57 crc kubenswrapper[4735]: W0215 20:32:57.437675 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda01ba16b_aa0c_4811_8223_9f4c083fdeab.slice/crio-9af9aae7e9d8e0991e8aa19dedfab167f8f3677103db2c44de93bfd6690e5a60 WatchSource:0}: Error finding container 9af9aae7e9d8e0991e8aa19dedfab167f8f3677103db2c44de93bfd6690e5a60: Status 404 returned error can't find the container with id 9af9aae7e9d8e0991e8aa19dedfab167f8f3677103db2c44de93bfd6690e5a60 Feb 15 20:32:57 crc kubenswrapper[4735]: I0215 20:32:57.484495 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a01ba16b-aa0c-4811-8223-9f4c083fdeab","Type":"ContainerStarted","Data":"9af9aae7e9d8e0991e8aa19dedfab167f8f3677103db2c44de93bfd6690e5a60"} Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.146482 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tk256"] Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.148709 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.152833 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.153018 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-nxdz8" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.160165 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.161704 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tk256"] Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.217241 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-fgcfc"] Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.220686 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.234251 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fgcfc"] Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.298814 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxbwk\" (UniqueName: \"kubernetes.io/projected/df76a192-fa00-4146-8f3a-8f7fa6798fec-kube-api-access-dxbwk\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.298864 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/df76a192-fa00-4146-8f3a-8f7fa6798fec-var-run-ovn\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.298899 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df76a192-fa00-4146-8f3a-8f7fa6798fec-combined-ca-bundle\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.298954 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df76a192-fa00-4146-8f3a-8f7fa6798fec-scripts\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.298978 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df76a192-fa00-4146-8f3a-8f7fa6798fec-var-run\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.299003 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/df76a192-fa00-4146-8f3a-8f7fa6798fec-ovn-controller-tls-certs\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.299026 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/df76a192-fa00-4146-8f3a-8f7fa6798fec-var-log-ovn\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402027 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/df76a192-fa00-4146-8f3a-8f7fa6798fec-var-log-ovn\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402095 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxbwk\" (UniqueName: \"kubernetes.io/projected/df76a192-fa00-4146-8f3a-8f7fa6798fec-kube-api-access-dxbwk\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402122 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/df76a192-fa00-4146-8f3a-8f7fa6798fec-var-run-ovn\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402142 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s475r\" (UniqueName: \"kubernetes.io/projected/9ca538ce-5609-4a07-887c-b9757cd7ad37-kube-api-access-s475r\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402174 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ca538ce-5609-4a07-887c-b9757cd7ad37-scripts\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402201 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df76a192-fa00-4146-8f3a-8f7fa6798fec-combined-ca-bundle\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402216 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-var-lib\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402246 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-etc-ovs\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402278 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-var-run\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402302 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df76a192-fa00-4146-8f3a-8f7fa6798fec-scripts\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df76a192-fa00-4146-8f3a-8f7fa6798fec-var-run\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402351 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-var-log\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402402 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/df76a192-fa00-4146-8f3a-8f7fa6798fec-ovn-controller-tls-certs\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.402999 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/df76a192-fa00-4146-8f3a-8f7fa6798fec-var-run-ovn\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.403165 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/df76a192-fa00-4146-8f3a-8f7fa6798fec-var-log-ovn\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.404325 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df76a192-fa00-4146-8f3a-8f7fa6798fec-var-run\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.412180 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df76a192-fa00-4146-8f3a-8f7fa6798fec-scripts\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.431823 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df76a192-fa00-4146-8f3a-8f7fa6798fec-combined-ca-bundle\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.437736 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxbwk\" (UniqueName: \"kubernetes.io/projected/df76a192-fa00-4146-8f3a-8f7fa6798fec-kube-api-access-dxbwk\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.444286 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/df76a192-fa00-4146-8f3a-8f7fa6798fec-ovn-controller-tls-certs\") pod \"ovn-controller-tk256\" (UID: \"df76a192-fa00-4146-8f3a-8f7fa6798fec\") " pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.495972 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.505861 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-etc-ovs\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.505905 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-var-run\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.505988 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-var-log\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.506077 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s475r\" (UniqueName: \"kubernetes.io/projected/9ca538ce-5609-4a07-887c-b9757cd7ad37-kube-api-access-s475r\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.506105 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ca538ce-5609-4a07-887c-b9757cd7ad37-scripts\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.506119 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-etc-ovs\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.506180 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-var-lib\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.507120 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-var-lib\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.521052 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ca538ce-5609-4a07-887c-b9757cd7ad37-scripts\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.522749 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-var-log\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.523125 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ca538ce-5609-4a07-887c-b9757cd7ad37-var-run\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.532653 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s475r\" (UniqueName: \"kubernetes.io/projected/9ca538ce-5609-4a07-887c-b9757cd7ad37-kube-api-access-s475r\") pod \"ovn-controller-ovs-fgcfc\" (UID: \"9ca538ce-5609-4a07-887c-b9757cd7ad37\") " pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.555005 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.931759 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.932992 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.937746 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.938540 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.938732 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.939009 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-mmxc7" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.939100 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 15 20:32:59 crc kubenswrapper[4735]: I0215 20:32:59.941029 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.121079 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.121148 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.121277 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.121351 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.121395 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.121410 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-config\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.121463 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.121521 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gcnl\" (UniqueName: \"kubernetes.io/projected/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-kube-api-access-2gcnl\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.223403 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.223446 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.223466 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-config\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.223495 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.223524 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gcnl\" (UniqueName: \"kubernetes.io/projected/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-kube-api-access-2gcnl\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.223549 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.223593 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.223625 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.225564 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.225900 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.232126 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.233066 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.246330 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.250298 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-config\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.273040 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gcnl\" (UniqueName: \"kubernetes.io/projected/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-kube-api-access-2gcnl\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.287148 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/77c1f16a-92e0-4c7a-a92c-12328f59f0b8-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.300246 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"77c1f16a-92e0-4c7a-a92c-12328f59f0b8\") " pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:00 crc kubenswrapper[4735]: I0215 20:33:00.571858 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.807655 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.809052 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.811051 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.811366 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-9vds2" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.811546 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.811716 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.834148 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.917500 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.917576 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.917597 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.917671 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-config\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.917710 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwr9t\" (UniqueName: \"kubernetes.io/projected/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-kube-api-access-fwr9t\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.917743 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.917763 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:02 crc kubenswrapper[4735]: I0215 20:33:02.917790 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.020599 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.020646 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.020672 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.020699 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-config\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.020762 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwr9t\" (UniqueName: \"kubernetes.io/projected/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-kube-api-access-fwr9t\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.020807 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.020831 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.020873 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.021607 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.021753 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.021785 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.023337 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-config\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.033975 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.037780 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.039042 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.059121 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwr9t\" (UniqueName: \"kubernetes.io/projected/8bfd475e-d45a-4d6e-8150-90da2bf08b6a-kube-api-access-fwr9t\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.082872 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8bfd475e-d45a-4d6e-8150-90da2bf08b6a\") " pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:03 crc kubenswrapper[4735]: I0215 20:33:03.142570 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:12 crc kubenswrapper[4735]: E0215 20:33:12.522749 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Feb 15 20:33:12 crc kubenswrapper[4735]: E0215 20:33:12.523531 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-svvvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(aa84f1ba-392a-4523-927c-405fab5bd619): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:33:12 crc kubenswrapper[4735]: E0215 20:33:12.525084 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="aa84f1ba-392a-4523-927c-405fab5bd619" Feb 15 20:33:12 crc kubenswrapper[4735]: E0215 20:33:12.638337 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="aa84f1ba-392a-4523-927c-405fab5bd619" Feb 15 20:33:14 crc kubenswrapper[4735]: E0215 20:33:14.592127 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Feb 15 20:33:14 crc kubenswrapper[4735]: E0215 20:33:14.592497 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zbxfc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(1f87d118-86d0-477e-a735-d2cd281b334a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:33:14 crc kubenswrapper[4735]: E0215 20:33:14.593999 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="1f87d118-86d0-477e-a735-d2cd281b334a" Feb 15 20:33:14 crc kubenswrapper[4735]: E0215 20:33:14.649490 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="1f87d118-86d0-477e-a735-d2cd281b334a" Feb 15 20:33:16 crc kubenswrapper[4735]: E0215 20:33:16.573432 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/openstack-rabbitmq:r42p" Feb 15 20:33:16 crc kubenswrapper[4735]: E0215 20:33:16.573722 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/lmiccini/openstack-rabbitmq:r42p,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-52cqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(dc3c285a-910a-4cb1-968e-636216557d09): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:33:16 crc kubenswrapper[4735]: E0215 20:33:16.574953 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="dc3c285a-910a-4cb1-968e-636216557d09" Feb 15 20:33:16 crc kubenswrapper[4735]: E0215 20:33:16.595389 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/openstack-rabbitmq:r42p" Feb 15 20:33:16 crc kubenswrapper[4735]: E0215 20:33:16.595576 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/lmiccini/openstack-rabbitmq:r42p,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9727r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(5c7b91a3-10d6-465d-aff4-4855d697b54f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:33:16 crc kubenswrapper[4735]: E0215 20:33:16.596754 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="5c7b91a3-10d6-465d-aff4-4855d697b54f" Feb 15 20:33:16 crc kubenswrapper[4735]: E0215 20:33:16.662041 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/openstack-rabbitmq:r42p\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="5c7b91a3-10d6-465d-aff4-4855d697b54f" Feb 15 20:33:16 crc kubenswrapper[4735]: E0215 20:33:16.664496 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/openstack-rabbitmq:r42p\\\"\"" pod="openstack/rabbitmq-server-0" podUID="dc3c285a-910a-4cb1-968e-636216557d09" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.435800 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.438182 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b8fvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-d4n6w_openstack(49cf9284-c3f4-44a6-8625-2fecc515065a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.440269 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" podUID="49cf9284-c3f4-44a6-8625-2fecc515065a" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.448449 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.448682 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6c2ld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-drgd2_openstack(1dfdacbc-8183-4181-854b-bb0be9230eb8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.449258 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.449573 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5qhds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-lz8md_openstack(587df3d0-fd47-4cc8-8d59-bd9a34c19a79): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.450877 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" podUID="587df3d0-fd47-4cc8-8d59-bd9a34c19a79" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.450960 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-drgd2" podUID="1dfdacbc-8183-4181-854b-bb0be9230eb8" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.505413 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.505607 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4mlbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-jzftf_openstack(d5382c75-62d2-412c-ab22-e1a8a83a2e71): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.507879 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" podUID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" Feb 15 20:33:22 crc kubenswrapper[4735]: I0215 20:33:22.702095 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6a34b649-4566-4a07-846e-e97886eec195","Type":"ContainerStarted","Data":"e64f9e12c2d46cc006185b8fba4d0d8461f080a31049c08267f5695f9ff9386e"} Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.708526 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-drgd2" podUID="1dfdacbc-8183-4181-854b-bb0be9230eb8" Feb 15 20:33:22 crc kubenswrapper[4735]: E0215 20:33:22.708623 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" podUID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" Feb 15 20:33:22 crc kubenswrapper[4735]: I0215 20:33:22.819727 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.934377873 podStartE2EDuration="29.819704746s" podCreationTimestamp="2026-02-15 20:32:53 +0000 UTC" firstStartedPulling="2026-02-15 20:32:54.85243127 +0000 UTC m=+982.718446893" lastFinishedPulling="2026-02-15 20:33:20.737758113 +0000 UTC m=+1008.603773766" observedRunningTime="2026-02-15 20:33:22.802844913 +0000 UTC m=+1010.668860536" watchObservedRunningTime="2026-02-15 20:33:22.819704746 +0000 UTC m=+1010.685720369" Feb 15 20:33:22 crc kubenswrapper[4735]: I0215 20:33:22.864832 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 15 20:33:22 crc kubenswrapper[4735]: I0215 20:33:22.975304 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tk256"] Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.219772 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fgcfc"] Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.296789 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.363627 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.369141 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.476180 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8fvn\" (UniqueName: \"kubernetes.io/projected/49cf9284-c3f4-44a6-8625-2fecc515065a-kube-api-access-b8fvn\") pod \"49cf9284-c3f4-44a6-8625-2fecc515065a\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.477531 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-config\") pod \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\" (UID: \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\") " Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.477566 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-dns-svc\") pod \"49cf9284-c3f4-44a6-8625-2fecc515065a\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.477786 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-config\") pod \"49cf9284-c3f4-44a6-8625-2fecc515065a\" (UID: \"49cf9284-c3f4-44a6-8625-2fecc515065a\") " Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.477920 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qhds\" (UniqueName: \"kubernetes.io/projected/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-kube-api-access-5qhds\") pod \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\" (UID: \"587df3d0-fd47-4cc8-8d59-bd9a34c19a79\") " Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.478449 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-config" (OuterVolumeSpecName: "config") pod "49cf9284-c3f4-44a6-8625-2fecc515065a" (UID: "49cf9284-c3f4-44a6-8625-2fecc515065a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.478485 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-config" (OuterVolumeSpecName: "config") pod "587df3d0-fd47-4cc8-8d59-bd9a34c19a79" (UID: "587df3d0-fd47-4cc8-8d59-bd9a34c19a79"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.478735 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.478752 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.479014 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "49cf9284-c3f4-44a6-8625-2fecc515065a" (UID: "49cf9284-c3f4-44a6-8625-2fecc515065a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.482596 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49cf9284-c3f4-44a6-8625-2fecc515065a-kube-api-access-b8fvn" (OuterVolumeSpecName: "kube-api-access-b8fvn") pod "49cf9284-c3f4-44a6-8625-2fecc515065a" (UID: "49cf9284-c3f4-44a6-8625-2fecc515065a"). InnerVolumeSpecName "kube-api-access-b8fvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.487646 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-kube-api-access-5qhds" (OuterVolumeSpecName: "kube-api-access-5qhds") pod "587df3d0-fd47-4cc8-8d59-bd9a34c19a79" (UID: "587df3d0-fd47-4cc8-8d59-bd9a34c19a79"). InnerVolumeSpecName "kube-api-access-5qhds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.579684 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf9284-c3f4-44a6-8625-2fecc515065a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.579716 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qhds\" (UniqueName: \"kubernetes.io/projected/587df3d0-fd47-4cc8-8d59-bd9a34c19a79-kube-api-access-5qhds\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.579726 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8fvn\" (UniqueName: \"kubernetes.io/projected/49cf9284-c3f4-44a6-8625-2fecc515065a-kube-api-access-b8fvn\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.711638 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.711638 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-d4n6w" event={"ID":"49cf9284-c3f4-44a6-8625-2fecc515065a","Type":"ContainerDied","Data":"9811f7ac0749b95d4332649c486e6d30baada140669b9cdc665c457f52e3cbe5"} Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.712718 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8bfd475e-d45a-4d6e-8150-90da2bf08b6a","Type":"ContainerStarted","Data":"b0714bfe0be1e2efe39b37633dd1c89de5c4cb282f71a610ed475da7ba8e53c4"} Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.716397 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tk256" event={"ID":"df76a192-fa00-4146-8f3a-8f7fa6798fec","Type":"ContainerStarted","Data":"751941593bce3c485ba276e6e4358f957bfc4310db4c079200b96877e45624a3"} Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.717487 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" event={"ID":"587df3d0-fd47-4cc8-8d59-bd9a34c19a79","Type":"ContainerDied","Data":"60613e892c1d79212af45d0e5f7840a591f3cbcf28c02f186430bdcd26a136d1"} Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.717564 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lz8md" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.721513 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"77c1f16a-92e0-4c7a-a92c-12328f59f0b8","Type":"ContainerStarted","Data":"b05410dea356f319a1fbab4d6389836786815815f8168a0a025aa2dda87a80b6"} Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.724430 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fgcfc" event={"ID":"9ca538ce-5609-4a07-887c-b9757cd7ad37","Type":"ContainerStarted","Data":"fae81078841663ed8167f3a67498ade45b5a2063d0242a299b5ec0793e848602"} Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.724616 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.812694 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d4n6w"] Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.827073 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d4n6w"] Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.836324 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lz8md"] Feb 15 20:33:23 crc kubenswrapper[4735]: I0215 20:33:23.840926 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lz8md"] Feb 15 20:33:24 crc kubenswrapper[4735]: I0215 20:33:24.734678 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a01ba16b-aa0c-4811-8223-9f4c083fdeab","Type":"ContainerStarted","Data":"1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec"} Feb 15 20:33:24 crc kubenswrapper[4735]: I0215 20:33:24.735031 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 15 20:33:24 crc kubenswrapper[4735]: I0215 20:33:24.755799 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.75013128 podStartE2EDuration="28.755777553s" podCreationTimestamp="2026-02-15 20:32:56 +0000 UTC" firstStartedPulling="2026-02-15 20:32:57.444808538 +0000 UTC m=+985.310824161" lastFinishedPulling="2026-02-15 20:33:24.450454811 +0000 UTC m=+1012.316470434" observedRunningTime="2026-02-15 20:33:24.747252915 +0000 UTC m=+1012.613268538" watchObservedRunningTime="2026-02-15 20:33:24.755777553 +0000 UTC m=+1012.621793176" Feb 15 20:33:24 crc kubenswrapper[4735]: I0215 20:33:24.894976 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49cf9284-c3f4-44a6-8625-2fecc515065a" path="/var/lib/kubelet/pods/49cf9284-c3f4-44a6-8625-2fecc515065a/volumes" Feb 15 20:33:24 crc kubenswrapper[4735]: I0215 20:33:24.895359 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="587df3d0-fd47-4cc8-8d59-bd9a34c19a79" path="/var/lib/kubelet/pods/587df3d0-fd47-4cc8-8d59-bd9a34c19a79/volumes" Feb 15 20:33:29 crc kubenswrapper[4735]: I0215 20:33:29.230758 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.448622 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kjhzm"] Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.450625 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.459535 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kjhzm"] Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.521842 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p54f\" (UniqueName: \"kubernetes.io/projected/db12b6bf-810b-44b6-b498-5550c7aaf23d-kube-api-access-4p54f\") pod \"certified-operators-kjhzm\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.522034 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-utilities\") pod \"certified-operators-kjhzm\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.522066 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-catalog-content\") pod \"certified-operators-kjhzm\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.623399 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p54f\" (UniqueName: \"kubernetes.io/projected/db12b6bf-810b-44b6-b498-5550c7aaf23d-kube-api-access-4p54f\") pod \"certified-operators-kjhzm\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.623573 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-utilities\") pod \"certified-operators-kjhzm\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.623598 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-catalog-content\") pod \"certified-operators-kjhzm\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.624114 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-catalog-content\") pod \"certified-operators-kjhzm\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.624344 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-utilities\") pod \"certified-operators-kjhzm\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.654593 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p54f\" (UniqueName: \"kubernetes.io/projected/db12b6bf-810b-44b6-b498-5550c7aaf23d-kube-api-access-4p54f\") pod \"certified-operators-kjhzm\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:30 crc kubenswrapper[4735]: I0215 20:33:30.797806 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.445601 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kjhzm"] Feb 15 20:33:31 crc kubenswrapper[4735]: W0215 20:33:31.452576 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb12b6bf_810b_44b6_b498_5550c7aaf23d.slice/crio-f1fe1df17d6bf62067d0d48c2f63ab1d344b662db355a30978e08cc0c8245efb WatchSource:0}: Error finding container f1fe1df17d6bf62067d0d48c2f63ab1d344b662db355a30978e08cc0c8245efb: Status 404 returned error can't find the container with id f1fe1df17d6bf62067d0d48c2f63ab1d344b662db355a30978e08cc0c8245efb Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.808861 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5c7b91a3-10d6-465d-aff4-4855d697b54f","Type":"ContainerStarted","Data":"528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca"} Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.810811 4735 generic.go:334] "Generic (PLEG): container finished" podID="9ca538ce-5609-4a07-887c-b9757cd7ad37" containerID="d9a5392d1a7b88dedbb711fbea68ea7226a6587616cf1652536464213413bddd" exitCode=0 Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.810845 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fgcfc" event={"ID":"9ca538ce-5609-4a07-887c-b9757cd7ad37","Type":"ContainerDied","Data":"d9a5392d1a7b88dedbb711fbea68ea7226a6587616cf1652536464213413bddd"} Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.812453 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8bfd475e-d45a-4d6e-8150-90da2bf08b6a","Type":"ContainerStarted","Data":"5a5741724d97d3be9f69bffcfb6a6ce3c98a24de3bd014d167ded43c6f7815d5"} Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.813608 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"aa84f1ba-392a-4523-927c-405fab5bd619","Type":"ContainerStarted","Data":"e725bd485f79634414e786b292b9612a50b9c6965f63b51aadb2760cfd702d1e"} Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.815358 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tk256" event={"ID":"df76a192-fa00-4146-8f3a-8f7fa6798fec","Type":"ContainerStarted","Data":"e9ed2f2044a50270d5aeff6b3df01d6f96e8b2cf89f058e7af459c6a2bec6b68"} Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.815443 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-tk256" Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.817967 4735 generic.go:334] "Generic (PLEG): container finished" podID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerID="6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8" exitCode=0 Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.818005 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjhzm" event={"ID":"db12b6bf-810b-44b6-b498-5550c7aaf23d","Type":"ContainerDied","Data":"6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8"} Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.818018 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjhzm" event={"ID":"db12b6bf-810b-44b6-b498-5550c7aaf23d","Type":"ContainerStarted","Data":"f1fe1df17d6bf62067d0d48c2f63ab1d344b662db355a30978e08cc0c8245efb"} Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.824186 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"77c1f16a-92e0-4c7a-a92c-12328f59f0b8","Type":"ContainerStarted","Data":"8268ab932ce118cba15336598c50e1febfe4da84db127223d0eb263c74241136"} Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.825931 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1f87d118-86d0-477e-a735-d2cd281b334a","Type":"ContainerStarted","Data":"8765ad89660a4317ed4936fbfff27d9ade969b3cb99373973f92791d02912406"} Feb 15 20:33:31 crc kubenswrapper[4735]: I0215 20:33:31.859517 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-tk256" podStartSLOduration=25.416050269 podStartE2EDuration="32.859501897s" podCreationTimestamp="2026-02-15 20:32:59 +0000 UTC" firstStartedPulling="2026-02-15 20:33:23.302805758 +0000 UTC m=+1011.168821371" lastFinishedPulling="2026-02-15 20:33:30.746257376 +0000 UTC m=+1018.612272999" observedRunningTime="2026-02-15 20:33:31.856722712 +0000 UTC m=+1019.722738335" watchObservedRunningTime="2026-02-15 20:33:31.859501897 +0000 UTC m=+1019.725517520" Feb 15 20:33:32 crc kubenswrapper[4735]: I0215 20:33:32.837221 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc3c285a-910a-4cb1-968e-636216557d09","Type":"ContainerStarted","Data":"6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a"} Feb 15 20:33:32 crc kubenswrapper[4735]: I0215 20:33:32.842636 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fgcfc" event={"ID":"9ca538ce-5609-4a07-887c-b9757cd7ad37","Type":"ContainerStarted","Data":"bbadfb6b52546fe1b18496783efd0a08b911fc9d196894fec3596994775c6e99"} Feb 15 20:33:32 crc kubenswrapper[4735]: I0215 20:33:32.842678 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fgcfc" event={"ID":"9ca538ce-5609-4a07-887c-b9757cd7ad37","Type":"ContainerStarted","Data":"6694c737507c235845b553b19743a5913804adfdb60dc36f9ad65a7c808861ae"} Feb 15 20:33:32 crc kubenswrapper[4735]: I0215 20:33:32.842756 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:33:32 crc kubenswrapper[4735]: I0215 20:33:32.842890 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:33:32 crc kubenswrapper[4735]: I0215 20:33:32.911263 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-fgcfc" podStartSLOduration=26.512326233 podStartE2EDuration="33.911235466s" podCreationTimestamp="2026-02-15 20:32:59 +0000 UTC" firstStartedPulling="2026-02-15 20:33:23.321727196 +0000 UTC m=+1011.187742819" lastFinishedPulling="2026-02-15 20:33:30.720636429 +0000 UTC m=+1018.586652052" observedRunningTime="2026-02-15 20:33:32.903526979 +0000 UTC m=+1020.769542612" watchObservedRunningTime="2026-02-15 20:33:32.911235466 +0000 UTC m=+1020.777251099" Feb 15 20:33:34 crc kubenswrapper[4735]: I0215 20:33:34.866004 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"77c1f16a-92e0-4c7a-a92c-12328f59f0b8","Type":"ContainerStarted","Data":"864e6416d9999c1efba2318dd1dc37b8a573ff57991dab929226351fd916446a"} Feb 15 20:33:34 crc kubenswrapper[4735]: I0215 20:33:34.870656 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8bfd475e-d45a-4d6e-8150-90da2bf08b6a","Type":"ContainerStarted","Data":"74117af0bb4f897f6bf1f482cd42044d98e7521fdc4e8fabe0bd6da3397ab6e1"} Feb 15 20:33:34 crc kubenswrapper[4735]: I0215 20:33:34.873336 4735 generic.go:334] "Generic (PLEG): container finished" podID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerID="c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b" exitCode=0 Feb 15 20:33:34 crc kubenswrapper[4735]: I0215 20:33:34.873424 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjhzm" event={"ID":"db12b6bf-810b-44b6-b498-5550c7aaf23d","Type":"ContainerDied","Data":"c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b"} Feb 15 20:33:34 crc kubenswrapper[4735]: I0215 20:33:34.900971 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=26.409270686 podStartE2EDuration="36.900921962s" podCreationTimestamp="2026-02-15 20:32:58 +0000 UTC" firstStartedPulling="2026-02-15 20:33:23.314673326 +0000 UTC m=+1011.180688939" lastFinishedPulling="2026-02-15 20:33:33.806324592 +0000 UTC m=+1021.672340215" observedRunningTime="2026-02-15 20:33:34.900289116 +0000 UTC m=+1022.766304749" watchObservedRunningTime="2026-02-15 20:33:34.900921962 +0000 UTC m=+1022.766937635" Feb 15 20:33:34 crc kubenswrapper[4735]: I0215 20:33:34.982565 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=23.550198717 podStartE2EDuration="33.982538102s" podCreationTimestamp="2026-02-15 20:33:01 +0000 UTC" firstStartedPulling="2026-02-15 20:33:23.315817377 +0000 UTC m=+1011.181833000" lastFinishedPulling="2026-02-15 20:33:33.748156762 +0000 UTC m=+1021.614172385" observedRunningTime="2026-02-15 20:33:34.969288287 +0000 UTC m=+1022.835303930" watchObservedRunningTime="2026-02-15 20:33:34.982538102 +0000 UTC m=+1022.848553745" Feb 15 20:33:35 crc kubenswrapper[4735]: I0215 20:33:35.572297 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:35 crc kubenswrapper[4735]: I0215 20:33:35.881327 4735 generic.go:334] "Generic (PLEG): container finished" podID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" containerID="8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280" exitCode=0 Feb 15 20:33:35 crc kubenswrapper[4735]: I0215 20:33:35.881400 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" event={"ID":"d5382c75-62d2-412c-ab22-e1a8a83a2e71","Type":"ContainerDied","Data":"8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280"} Feb 15 20:33:35 crc kubenswrapper[4735]: I0215 20:33:35.884113 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjhzm" event={"ID":"db12b6bf-810b-44b6-b498-5550c7aaf23d","Type":"ContainerStarted","Data":"b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96"} Feb 15 20:33:35 crc kubenswrapper[4735]: I0215 20:33:35.929965 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kjhzm" podStartSLOduration=2.30093093 podStartE2EDuration="5.929932283s" podCreationTimestamp="2026-02-15 20:33:30 +0000 UTC" firstStartedPulling="2026-02-15 20:33:31.819854962 +0000 UTC m=+1019.685870585" lastFinishedPulling="2026-02-15 20:33:35.448856315 +0000 UTC m=+1023.314871938" observedRunningTime="2026-02-15 20:33:35.924222109 +0000 UTC m=+1023.790237732" watchObservedRunningTime="2026-02-15 20:33:35.929932283 +0000 UTC m=+1023.795947906" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.143528 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.180758 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.572985 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.613728 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.616732 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jzftf"] Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.657257 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-bc82t"] Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.658740 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.682422 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-bc82t"] Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.738510 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb845\" (UniqueName: \"kubernetes.io/projected/32316c0b-9b63-48ba-a5b1-eacc93528c6e-kube-api-access-tb845\") pod \"dnsmasq-dns-7cb5889db5-bc82t\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.738566 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-bc82t\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.738602 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-config\") pod \"dnsmasq-dns-7cb5889db5-bc82t\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.811581 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.839716 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb845\" (UniqueName: \"kubernetes.io/projected/32316c0b-9b63-48ba-a5b1-eacc93528c6e-kube-api-access-tb845\") pod \"dnsmasq-dns-7cb5889db5-bc82t\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.839771 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-bc82t\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.839811 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-config\") pod \"dnsmasq-dns-7cb5889db5-bc82t\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.840701 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-bc82t\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.840710 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-config\") pod \"dnsmasq-dns-7cb5889db5-bc82t\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.871903 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb845\" (UniqueName: \"kubernetes.io/projected/32316c0b-9b63-48ba-a5b1-eacc93528c6e-kube-api-access-tb845\") pod \"dnsmasq-dns-7cb5889db5-bc82t\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.893675 4735 generic.go:334] "Generic (PLEG): container finished" podID="1f87d118-86d0-477e-a735-d2cd281b334a" containerID="8765ad89660a4317ed4936fbfff27d9ade969b3cb99373973f92791d02912406" exitCode=0 Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.896220 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1f87d118-86d0-477e-a735-d2cd281b334a","Type":"ContainerDied","Data":"8765ad89660a4317ed4936fbfff27d9ade969b3cb99373973f92791d02912406"} Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.898213 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" event={"ID":"d5382c75-62d2-412c-ab22-e1a8a83a2e71","Type":"ContainerStarted","Data":"557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633"} Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.898365 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" podUID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" containerName="dnsmasq-dns" containerID="cri-o://557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633" gracePeriod=10 Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.898628 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.903055 4735 generic.go:334] "Generic (PLEG): container finished" podID="aa84f1ba-392a-4523-927c-405fab5bd619" containerID="e725bd485f79634414e786b292b9612a50b9c6965f63b51aadb2760cfd702d1e" exitCode=0 Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.904163 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"aa84f1ba-392a-4523-927c-405fab5bd619","Type":"ContainerDied","Data":"e725bd485f79634414e786b292b9612a50b9c6965f63b51aadb2760cfd702d1e"} Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.904195 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.949376 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" podStartSLOduration=3.736786974 podStartE2EDuration="47.949361105s" podCreationTimestamp="2026-02-15 20:32:49 +0000 UTC" firstStartedPulling="2026-02-15 20:32:50.881182566 +0000 UTC m=+978.747198189" lastFinishedPulling="2026-02-15 20:33:35.093756657 +0000 UTC m=+1022.959772320" observedRunningTime="2026-02-15 20:33:36.947347872 +0000 UTC m=+1024.813363495" watchObservedRunningTime="2026-02-15 20:33:36.949361105 +0000 UTC m=+1024.815376728" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.980978 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.983715 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 15 20:33:36 crc kubenswrapper[4735]: I0215 20:33:36.988812 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.203908 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-drgd2"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.280830 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-62khd"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.282175 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.312172 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.346642 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-8vw9j"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.359891 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.379192 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.382572 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-dns-svc\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.382626 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4brwf\" (UniqueName: \"kubernetes.io/projected/2d4946ef-28a3-4831-adaf-ec2904168e44-kube-api-access-4brwf\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.382658 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.382774 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-config\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.439047 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8vw9j"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.467032 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-62khd"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.504887 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-dns-svc\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.504929 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4brwf\" (UniqueName: \"kubernetes.io/projected/2d4946ef-28a3-4831-adaf-ec2904168e44-kube-api-access-4brwf\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.504975 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3ff080f2-642e-4ff4-9554-5489daca64c4-ovn-rundir\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.504991 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.505015 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3ff080f2-642e-4ff4-9554-5489daca64c4-ovs-rundir\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.505032 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ff080f2-642e-4ff4-9554-5489daca64c4-config\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.506147 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-dns-svc\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.506153 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.506817 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-config\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.505068 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-config\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.516314 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pnw9\" (UniqueName: \"kubernetes.io/projected/3ff080f2-642e-4ff4-9554-5489daca64c4-kube-api-access-8pnw9\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.516339 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff080f2-642e-4ff4-9554-5489daca64c4-combined-ca-bundle\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.516398 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ff080f2-642e-4ff4-9554-5489daca64c4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.534912 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-bc82t"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.535512 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.541381 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4brwf\" (UniqueName: \"kubernetes.io/projected/2d4946ef-28a3-4831-adaf-ec2904168e44-kube-api-access-4brwf\") pod \"dnsmasq-dns-57d65f699f-62khd\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.541590 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.547771 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.547813 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.547858 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.548250 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-nvmlc" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.548838 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-k7qln"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.550266 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.565149 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.568303 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.575934 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-k7qln"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627305 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627341 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-config\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627360 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627390 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627422 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-config\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627441 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627460 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3ff080f2-642e-4ff4-9554-5489daca64c4-ovn-rundir\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627484 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5wtv\" (UniqueName: \"kubernetes.io/projected/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-kube-api-access-g5wtv\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627518 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3ff080f2-642e-4ff4-9554-5489daca64c4-ovs-rundir\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627537 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-scripts\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627555 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ff080f2-642e-4ff4-9554-5489daca64c4-config\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627591 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627677 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pnw9\" (UniqueName: \"kubernetes.io/projected/3ff080f2-642e-4ff4-9554-5489daca64c4-kube-api-access-8pnw9\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627693 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff080f2-642e-4ff4-9554-5489daca64c4-combined-ca-bundle\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627730 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ff080f2-642e-4ff4-9554-5489daca64c4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627761 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627795 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.627821 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkd58\" (UniqueName: \"kubernetes.io/projected/d482296c-b6fe-46a1-8798-696c4d06cdf0-kube-api-access-xkd58\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.628160 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3ff080f2-642e-4ff4-9554-5489daca64c4-ovn-rundir\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.628216 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3ff080f2-642e-4ff4-9554-5489daca64c4-ovs-rundir\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.628808 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ff080f2-642e-4ff4-9554-5489daca64c4-config\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.642549 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ff080f2-642e-4ff4-9554-5489daca64c4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.647048 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff080f2-642e-4ff4-9554-5489daca64c4-combined-ca-bundle\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.661576 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pnw9\" (UniqueName: \"kubernetes.io/projected/3ff080f2-642e-4ff4-9554-5489daca64c4-kube-api-access-8pnw9\") pod \"ovn-controller-metrics-8vw9j\" (UID: \"3ff080f2-642e-4ff4-9554-5489daca64c4\") " pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.694361 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.728721 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.728818 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-config\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.728845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.729664 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-config\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.729735 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5wtv\" (UniqueName: \"kubernetes.io/projected/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-kube-api-access-g5wtv\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.730093 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-scripts\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.732760 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-scripts\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.735073 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.736657 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.736830 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.737116 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.737162 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.737238 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkd58\" (UniqueName: \"kubernetes.io/projected/d482296c-b6fe-46a1-8798-696c4d06cdf0-kube-api-access-xkd58\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.738175 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.737252 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.738265 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-config\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.738097 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.738739 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.739377 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-config\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.740212 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.749758 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.756419 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkd58\" (UniqueName: \"kubernetes.io/projected/d482296c-b6fe-46a1-8798-696c4d06cdf0-kube-api-access-xkd58\") pod \"dnsmasq-dns-b8fbc5445-k7qln\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.756565 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.757378 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5wtv\" (UniqueName: \"kubernetes.io/projected/4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d-kube-api-access-g5wtv\") pod \"ovn-northd-0\" (UID: \"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d\") " pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.767182 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8vw9j" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.879253 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.887743 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.891117 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.900887 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.901270 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-jcthz" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.901406 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.901560 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.907277 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.915293 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.946530 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.946607 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-lock\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.946636 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.946664 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4skp8\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-kube-api-access-4skp8\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.946779 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.946809 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-cache\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:37 crc kubenswrapper[4735]: I0215 20:33:37.978816 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"aa84f1ba-392a-4523-927c-405fab5bd619","Type":"ContainerStarted","Data":"2f704bab9b2d66fff5f649b70578b57c84e399a3badb6c09cd819b05adaf967c"} Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.034555 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1f87d118-86d0-477e-a735-d2cd281b334a","Type":"ContainerStarted","Data":"52b4acf3e6c3243abdd108540698ef7969403c033a2a2b8c65cfa9eaf22a4e20"} Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.042415 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.046383 4735 generic.go:334] "Generic (PLEG): container finished" podID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" containerID="557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633" exitCode=0 Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.046561 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" event={"ID":"d5382c75-62d2-412c-ab22-e1a8a83a2e71","Type":"ContainerDied","Data":"557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633"} Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.046616 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" event={"ID":"d5382c75-62d2-412c-ab22-e1a8a83a2e71","Type":"ContainerDied","Data":"1d4bb0fec04c39257007a31c1763cdba4e62f479b113e32440eb8b61d9f4eff2"} Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.046639 4735 scope.go:117] "RemoveContainer" containerID="557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.048613 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-lock\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.048712 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.048790 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4skp8\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-kube-api-access-4skp8\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.048896 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.048983 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-cache\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.049172 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.049253 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: E0215 20:33:38.050352 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 15 20:33:38 crc kubenswrapper[4735]: E0215 20:33:38.050374 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 15 20:33:38 crc kubenswrapper[4735]: E0215 20:33:38.050408 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift podName:a9642a7f-e8aa-4368-b2b9-3da6537c4d8f nodeName:}" failed. No retries permitted until 2026-02-15 20:33:38.550393807 +0000 UTC m=+1026.416409430 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift") pod "swift-storage-0" (UID: "a9642a7f-e8aa-4368-b2b9-3da6537c4d8f") : configmap "swift-ring-files" not found Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.074671 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-bc82t"] Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.076858 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.683396214 podStartE2EDuration="47.076841227s" podCreationTimestamp="2026-02-15 20:32:51 +0000 UTC" firstStartedPulling="2026-02-15 20:32:53.385742667 +0000 UTC m=+981.251758290" lastFinishedPulling="2026-02-15 20:33:30.77918768 +0000 UTC m=+1018.645203303" observedRunningTime="2026-02-15 20:33:38.028560402 +0000 UTC m=+1025.894576025" watchObservedRunningTime="2026-02-15 20:33:38.076841227 +0000 UTC m=+1025.942856850" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.078035 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-lock\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.078310 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-cache\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.079077 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.085799 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4skp8\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-kube-api-access-4skp8\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.095305 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=9.859619003 podStartE2EDuration="46.095288082s" podCreationTimestamp="2026-02-15 20:32:52 +0000 UTC" firstStartedPulling="2026-02-15 20:32:54.54758555 +0000 UTC m=+982.413601173" lastFinishedPulling="2026-02-15 20:33:30.783254629 +0000 UTC m=+1018.649270252" observedRunningTime="2026-02-15 20:33:38.07171184 +0000 UTC m=+1025.937727463" watchObservedRunningTime="2026-02-15 20:33:38.095288082 +0000 UTC m=+1025.961303705" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.102183 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.141410 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.147549 4735 scope.go:117] "RemoveContainer" containerID="8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.257550 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-config\") pod \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.257928 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mlbv\" (UniqueName: \"kubernetes.io/projected/d5382c75-62d2-412c-ab22-e1a8a83a2e71-kube-api-access-4mlbv\") pod \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.258048 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-config\") pod \"1dfdacbc-8183-4181-854b-bb0be9230eb8\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.267376 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-config" (OuterVolumeSpecName: "config") pod "1dfdacbc-8183-4181-854b-bb0be9230eb8" (UID: "1dfdacbc-8183-4181-854b-bb0be9230eb8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.268162 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c2ld\" (UniqueName: \"kubernetes.io/projected/1dfdacbc-8183-4181-854b-bb0be9230eb8-kube-api-access-6c2ld\") pod \"1dfdacbc-8183-4181-854b-bb0be9230eb8\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.268224 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-dns-svc\") pod \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\" (UID: \"d5382c75-62d2-412c-ab22-e1a8a83a2e71\") " Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.268252 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-dns-svc\") pod \"1dfdacbc-8183-4181-854b-bb0be9230eb8\" (UID: \"1dfdacbc-8183-4181-854b-bb0be9230eb8\") " Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.273215 4735 scope.go:117] "RemoveContainer" containerID="557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633" Feb 15 20:33:38 crc kubenswrapper[4735]: E0215 20:33:38.276251 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633\": container with ID starting with 557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633 not found: ID does not exist" containerID="557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.276304 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633"} err="failed to get container status \"557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633\": rpc error: code = NotFound desc = could not find container \"557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633\": container with ID starting with 557e266b2e53204cf83a62dbb56149d2e7860620c55735439a7b65a225775633 not found: ID does not exist" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.276334 4735 scope.go:117] "RemoveContainer" containerID="8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280" Feb 15 20:33:38 crc kubenswrapper[4735]: E0215 20:33:38.278633 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280\": container with ID starting with 8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280 not found: ID does not exist" containerID="8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.278717 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280"} err="failed to get container status \"8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280\": rpc error: code = NotFound desc = could not find container \"8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280\": container with ID starting with 8870ef0133e333bed59c80a1adf13c7070e90e832ff4e7075684bd44ac6fa280 not found: ID does not exist" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.279283 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1dfdacbc-8183-4181-854b-bb0be9230eb8" (UID: "1dfdacbc-8183-4181-854b-bb0be9230eb8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.284906 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dfdacbc-8183-4181-854b-bb0be9230eb8-kube-api-access-6c2ld" (OuterVolumeSpecName: "kube-api-access-6c2ld") pod "1dfdacbc-8183-4181-854b-bb0be9230eb8" (UID: "1dfdacbc-8183-4181-854b-bb0be9230eb8"). InnerVolumeSpecName "kube-api-access-6c2ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.287126 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5382c75-62d2-412c-ab22-e1a8a83a2e71-kube-api-access-4mlbv" (OuterVolumeSpecName: "kube-api-access-4mlbv") pod "d5382c75-62d2-412c-ab22-e1a8a83a2e71" (UID: "d5382c75-62d2-412c-ab22-e1a8a83a2e71"). InnerVolumeSpecName "kube-api-access-4mlbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.287230 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.287247 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c2ld\" (UniqueName: \"kubernetes.io/projected/1dfdacbc-8183-4181-854b-bb0be9230eb8-kube-api-access-6c2ld\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.287265 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfdacbc-8183-4181-854b-bb0be9230eb8-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.335167 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-62khd"] Feb 15 20:33:38 crc kubenswrapper[4735]: W0215 20:33:38.349886 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d4946ef_28a3_4831_adaf_ec2904168e44.slice/crio-3f49d81c4fffa5a4410671237df5491df071b81101e63347132aa689ab7f7fec WatchSource:0}: Error finding container 3f49d81c4fffa5a4410671237df5491df071b81101e63347132aa689ab7f7fec: Status 404 returned error can't find the container with id 3f49d81c4fffa5a4410671237df5491df071b81101e63347132aa689ab7f7fec Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.363611 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-config" (OuterVolumeSpecName: "config") pod "d5382c75-62d2-412c-ab22-e1a8a83a2e71" (UID: "d5382c75-62d2-412c-ab22-e1a8a83a2e71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.369408 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5382c75-62d2-412c-ab22-e1a8a83a2e71" (UID: "d5382c75-62d2-412c-ab22-e1a8a83a2e71"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.389969 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.389999 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5382c75-62d2-412c-ab22-e1a8a83a2e71-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.390011 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mlbv\" (UniqueName: \"kubernetes.io/projected/d5382c75-62d2-412c-ab22-e1a8a83a2e71-kube-api-access-4mlbv\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.594174 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:38 crc kubenswrapper[4735]: E0215 20:33:38.594405 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 15 20:33:38 crc kubenswrapper[4735]: E0215 20:33:38.594420 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 15 20:33:38 crc kubenswrapper[4735]: E0215 20:33:38.594464 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift podName:a9642a7f-e8aa-4368-b2b9-3da6537c4d8f nodeName:}" failed. No retries permitted until 2026-02-15 20:33:39.594449786 +0000 UTC m=+1027.460465409 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift") pod "swift-storage-0" (UID: "a9642a7f-e8aa-4368-b2b9-3da6537c4d8f") : configmap "swift-ring-files" not found Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.601080 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8vw9j"] Feb 15 20:33:38 crc kubenswrapper[4735]: W0215 20:33:38.626638 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ff080f2_642e_4ff4_9554_5489daca64c4.slice/crio-c57fe2023db0db5785b1b76e774274996e61af813a66ada1003e1af7181bae99 WatchSource:0}: Error finding container c57fe2023db0db5785b1b76e774274996e61af813a66ada1003e1af7181bae99: Status 404 returned error can't find the container with id c57fe2023db0db5785b1b76e774274996e61af813a66ada1003e1af7181bae99 Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.731917 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 15 20:33:38 crc kubenswrapper[4735]: I0215 20:33:38.739969 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-k7qln"] Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.055176 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jzftf" Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.059478 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d","Type":"ContainerStarted","Data":"abefe4f86b80e9e36a8b183a43db7cef679f52f95c3e74a8af55402792dc36c1"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.062318 4735 generic.go:334] "Generic (PLEG): container finished" podID="d482296c-b6fe-46a1-8798-696c4d06cdf0" containerID="b8fa7ff7182371f53f58a31bc8ce267947d2a87a704042809da492950fc31605" exitCode=0 Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.062394 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" event={"ID":"d482296c-b6fe-46a1-8798-696c4d06cdf0","Type":"ContainerDied","Data":"b8fa7ff7182371f53f58a31bc8ce267947d2a87a704042809da492950fc31605"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.062417 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" event={"ID":"d482296c-b6fe-46a1-8798-696c4d06cdf0","Type":"ContainerStarted","Data":"44bf1115d9e7a2fb2408bdcefa715230b851b1df5625f771410b911c5ca95ebd"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.066589 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8vw9j" event={"ID":"3ff080f2-642e-4ff4-9554-5489daca64c4","Type":"ContainerStarted","Data":"cb250fb9d97c556fb5d6cb118aa58f496b04e8ed7446f9a1d50ef21a4c72d505"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.066635 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8vw9j" event={"ID":"3ff080f2-642e-4ff4-9554-5489daca64c4","Type":"ContainerStarted","Data":"c57fe2023db0db5785b1b76e774274996e61af813a66ada1003e1af7181bae99"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.070616 4735 generic.go:334] "Generic (PLEG): container finished" podID="2d4946ef-28a3-4831-adaf-ec2904168e44" containerID="8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975" exitCode=0 Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.070697 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-62khd" event={"ID":"2d4946ef-28a3-4831-adaf-ec2904168e44","Type":"ContainerDied","Data":"8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.070722 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-62khd" event={"ID":"2d4946ef-28a3-4831-adaf-ec2904168e44","Type":"ContainerStarted","Data":"3f49d81c4fffa5a4410671237df5491df071b81101e63347132aa689ab7f7fec"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.081907 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jzftf"] Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.082773 4735 generic.go:334] "Generic (PLEG): container finished" podID="32316c0b-9b63-48ba-a5b1-eacc93528c6e" containerID="eee031dd12369cb4ff41810078f7762b86242705428069b3425fde920f9fa7a3" exitCode=0 Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.082836 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" event={"ID":"32316c0b-9b63-48ba-a5b1-eacc93528c6e","Type":"ContainerDied","Data":"eee031dd12369cb4ff41810078f7762b86242705428069b3425fde920f9fa7a3"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.082863 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" event={"ID":"32316c0b-9b63-48ba-a5b1-eacc93528c6e","Type":"ContainerStarted","Data":"0b892732df45f6d3f57f9415b4016d0f2f0ec62f1718437f8dce9d8ce894738c"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.086540 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-drgd2" Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.086929 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-drgd2" event={"ID":"1dfdacbc-8183-4181-854b-bb0be9230eb8","Type":"ContainerDied","Data":"78d7757a973808b804921d881e142efa128ff6c1403eea17a7c63b53ec5864d7"} Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.091980 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jzftf"] Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.227670 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-8vw9j" podStartSLOduration=2.227654416 podStartE2EDuration="2.227654416s" podCreationTimestamp="2026-02-15 20:33:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:33:39.193282743 +0000 UTC m=+1027.059298366" watchObservedRunningTime="2026-02-15 20:33:39.227654416 +0000 UTC m=+1027.093670039" Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.314780 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-drgd2"] Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.320978 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-drgd2"] Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.514902 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.621659 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-config\") pod \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.621732 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb845\" (UniqueName: \"kubernetes.io/projected/32316c0b-9b63-48ba-a5b1-eacc93528c6e-kube-api-access-tb845\") pod \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.621903 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-dns-svc\") pod \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\" (UID: \"32316c0b-9b63-48ba-a5b1-eacc93528c6e\") " Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.622238 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:39 crc kubenswrapper[4735]: E0215 20:33:39.622429 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 15 20:33:39 crc kubenswrapper[4735]: E0215 20:33:39.622444 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 15 20:33:39 crc kubenswrapper[4735]: E0215 20:33:39.622495 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift podName:a9642a7f-e8aa-4368-b2b9-3da6537c4d8f nodeName:}" failed. No retries permitted until 2026-02-15 20:33:41.622477289 +0000 UTC m=+1029.488492922 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift") pod "swift-storage-0" (UID: "a9642a7f-e8aa-4368-b2b9-3da6537c4d8f") : configmap "swift-ring-files" not found Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.633407 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32316c0b-9b63-48ba-a5b1-eacc93528c6e-kube-api-access-tb845" (OuterVolumeSpecName: "kube-api-access-tb845") pod "32316c0b-9b63-48ba-a5b1-eacc93528c6e" (UID: "32316c0b-9b63-48ba-a5b1-eacc93528c6e"). InnerVolumeSpecName "kube-api-access-tb845". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.640998 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "32316c0b-9b63-48ba-a5b1-eacc93528c6e" (UID: "32316c0b-9b63-48ba-a5b1-eacc93528c6e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.646901 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-config" (OuterVolumeSpecName: "config") pod "32316c0b-9b63-48ba-a5b1-eacc93528c6e" (UID: "32316c0b-9b63-48ba-a5b1-eacc93528c6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.727387 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.727418 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32316c0b-9b63-48ba-a5b1-eacc93528c6e-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:39 crc kubenswrapper[4735]: I0215 20:33:39.727427 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb845\" (UniqueName: \"kubernetes.io/projected/32316c0b-9b63-48ba-a5b1-eacc93528c6e-kube-api-access-tb845\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.093196 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.093199 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-bc82t" event={"ID":"32316c0b-9b63-48ba-a5b1-eacc93528c6e","Type":"ContainerDied","Data":"0b892732df45f6d3f57f9415b4016d0f2f0ec62f1718437f8dce9d8ce894738c"} Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.093291 4735 scope.go:117] "RemoveContainer" containerID="eee031dd12369cb4ff41810078f7762b86242705428069b3425fde920f9fa7a3" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.096338 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" event={"ID":"d482296c-b6fe-46a1-8798-696c4d06cdf0","Type":"ContainerStarted","Data":"d8280a3ef47c7e342bea709f44e836795ab36a022326f5c00708516014e83e20"} Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.097228 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.100758 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-62khd" event={"ID":"2d4946ef-28a3-4831-adaf-ec2904168e44","Type":"ContainerStarted","Data":"9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9"} Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.100798 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.174778 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" podStartSLOduration=3.174761898 podStartE2EDuration="3.174761898s" podCreationTimestamp="2026-02-15 20:33:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:33:40.158101361 +0000 UTC m=+1028.024116984" watchObservedRunningTime="2026-02-15 20:33:40.174761898 +0000 UTC m=+1028.040777521" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.215165 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d65f699f-62khd" podStartSLOduration=3.215149981 podStartE2EDuration="3.215149981s" podCreationTimestamp="2026-02-15 20:33:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:33:40.175381314 +0000 UTC m=+1028.041396957" watchObservedRunningTime="2026-02-15 20:33:40.215149981 +0000 UTC m=+1028.081165604" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.252182 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-bc82t"] Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.260253 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-bc82t"] Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.798650 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.799875 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.846408 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.897580 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dfdacbc-8183-4181-854b-bb0be9230eb8" path="/var/lib/kubelet/pods/1dfdacbc-8183-4181-854b-bb0be9230eb8/volumes" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.898600 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32316c0b-9b63-48ba-a5b1-eacc93528c6e" path="/var/lib/kubelet/pods/32316c0b-9b63-48ba-a5b1-eacc93528c6e/volumes" Feb 15 20:33:40 crc kubenswrapper[4735]: I0215 20:33:40.899902 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" path="/var/lib/kubelet/pods/d5382c75-62d2-412c-ab22-e1a8a83a2e71/volumes" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.111197 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d","Type":"ContainerStarted","Data":"605e987ec2970fceb9ef0004559d4f9b433b3f71719493bad8dda778e73bc68c"} Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.111240 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d","Type":"ContainerStarted","Data":"4c5482d9281bf02e59c9c36c16b76087a87f0c8e7b5372a83e67a0dd5e12a020"} Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.112188 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.139664 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.829430932 podStartE2EDuration="4.139647256s" podCreationTimestamp="2026-02-15 20:33:37 +0000 UTC" firstStartedPulling="2026-02-15 20:33:38.753916745 +0000 UTC m=+1026.619932368" lastFinishedPulling="2026-02-15 20:33:40.064133069 +0000 UTC m=+1027.930148692" observedRunningTime="2026-02-15 20:33:41.134385355 +0000 UTC m=+1029.000400978" watchObservedRunningTime="2026-02-15 20:33:41.139647256 +0000 UTC m=+1029.005662879" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.173930 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.212535 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kjhzm"] Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.660228 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-4n5d2"] Feb 15 20:33:41 crc kubenswrapper[4735]: E0215 20:33:41.660589 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32316c0b-9b63-48ba-a5b1-eacc93528c6e" containerName="init" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.660608 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="32316c0b-9b63-48ba-a5b1-eacc93528c6e" containerName="init" Feb 15 20:33:41 crc kubenswrapper[4735]: E0215 20:33:41.660639 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" containerName="init" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.660648 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" containerName="init" Feb 15 20:33:41 crc kubenswrapper[4735]: E0215 20:33:41.660660 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" containerName="dnsmasq-dns" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.660670 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" containerName="dnsmasq-dns" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.660895 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="32316c0b-9b63-48ba-a5b1-eacc93528c6e" containerName="init" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.660918 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5382c75-62d2-412c-ab22-e1a8a83a2e71" containerName="dnsmasq-dns" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.661742 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.662507 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:41 crc kubenswrapper[4735]: E0215 20:33:41.662704 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 15 20:33:41 crc kubenswrapper[4735]: E0215 20:33:41.662747 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 15 20:33:41 crc kubenswrapper[4735]: E0215 20:33:41.662820 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift podName:a9642a7f-e8aa-4368-b2b9-3da6537c4d8f nodeName:}" failed. No retries permitted until 2026-02-15 20:33:45.662795974 +0000 UTC m=+1033.528811637 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift") pod "swift-storage-0" (UID: "a9642a7f-e8aa-4368-b2b9-3da6537c4d8f") : configmap "swift-ring-files" not found Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.677984 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.678075 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.678092 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.679226 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-4n5d2"] Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.730681 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-4n5d2"] Feb 15 20:33:41 crc kubenswrapper[4735]: E0215 20:33:41.731228 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-8vnrh ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-4n5d2" podUID="b063d1da-ed32-4c30-8b2a-2e65a4fcaceb" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.740243 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-zrmz5"] Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.741553 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.748905 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zrmz5"] Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.763999 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-scripts\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.764165 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vnrh\" (UniqueName: \"kubernetes.io/projected/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-kube-api-access-8vnrh\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.764222 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-ring-data-devices\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.764252 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-swiftconf\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.764295 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-etc-swift\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.764341 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-combined-ca-bundle\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.764397 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-dispersionconf\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.865768 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-swiftconf\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.865815 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-combined-ca-bundle\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.865847 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-dispersionconf\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.865974 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-scripts\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866067 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-dispersionconf\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866098 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-scripts\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866162 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jdrb\" (UniqueName: \"kubernetes.io/projected/35e018a0-8bf3-4725-a229-5a4196d4ae97-kube-api-access-8jdrb\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866182 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-combined-ca-bundle\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866215 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-ring-data-devices\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866248 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vnrh\" (UniqueName: \"kubernetes.io/projected/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-kube-api-access-8vnrh\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866308 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-ring-data-devices\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866340 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-swiftconf\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866360 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-etc-swift\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866402 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35e018a0-8bf3-4725-a229-5a4196d4ae97-etc-swift\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866752 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-scripts\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.866817 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-etc-swift\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.867331 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-ring-data-devices\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.870865 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-swiftconf\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.871447 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-combined-ca-bundle\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.882270 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-dispersionconf\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.895221 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vnrh\" (UniqueName: \"kubernetes.io/projected/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-kube-api-access-8vnrh\") pod \"swift-ring-rebalance-4n5d2\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.968081 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-scripts\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.968150 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-dispersionconf\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.968201 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jdrb\" (UniqueName: \"kubernetes.io/projected/35e018a0-8bf3-4725-a229-5a4196d4ae97-kube-api-access-8jdrb\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.968221 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-combined-ca-bundle\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.968248 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-ring-data-devices\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.968293 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35e018a0-8bf3-4725-a229-5a4196d4ae97-etc-swift\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.968316 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-swiftconf\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.969162 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-scripts\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.969505 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-ring-data-devices\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.969736 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35e018a0-8bf3-4725-a229-5a4196d4ae97-etc-swift\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.971043 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-swiftconf\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.973262 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-combined-ca-bundle\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.973644 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-dispersionconf\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:41 crc kubenswrapper[4735]: I0215 20:33:41.984373 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jdrb\" (UniqueName: \"kubernetes.io/projected/35e018a0-8bf3-4725-a229-5a4196d4ae97-kube-api-access-8jdrb\") pod \"swift-ring-rebalance-zrmz5\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.058883 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.128506 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.218465 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.375325 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-dispersionconf\") pod \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.375755 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vnrh\" (UniqueName: \"kubernetes.io/projected/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-kube-api-access-8vnrh\") pod \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.375803 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-swiftconf\") pod \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.375834 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-combined-ca-bundle\") pod \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.375907 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-scripts\") pod \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.375935 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-etc-swift\") pod \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.376000 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-ring-data-devices\") pod \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\" (UID: \"b063d1da-ed32-4c30-8b2a-2e65a4fcaceb\") " Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.376481 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-scripts" (OuterVolumeSpecName: "scripts") pod "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb" (UID: "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.376638 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb" (UID: "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.376729 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb" (UID: "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.380831 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb" (UID: "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.380872 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-kube-api-access-8vnrh" (OuterVolumeSpecName: "kube-api-access-8vnrh") pod "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb" (UID: "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb"). InnerVolumeSpecName "kube-api-access-8vnrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.381217 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb" (UID: "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.381610 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb" (UID: "b063d1da-ed32-4c30-8b2a-2e65a4fcaceb"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.477751 4735 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.477794 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vnrh\" (UniqueName: \"kubernetes.io/projected/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-kube-api-access-8vnrh\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.477811 4735 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.477822 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.477834 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.477845 4735 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.477856 4735 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.544008 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.544049 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.548126 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zrmz5"] Feb 15 20:33:42 crc kubenswrapper[4735]: W0215 20:33:42.552289 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35e018a0_8bf3_4725_a229_5a4196d4ae97.slice/crio-7ebf0a8fb4784c166a17ca4d7852e1ced8d53e9cb23f5593441274cae7ba041c WatchSource:0}: Error finding container 7ebf0a8fb4784c166a17ca4d7852e1ced8d53e9cb23f5593441274cae7ba041c: Status 404 returned error can't find the container with id 7ebf0a8fb4784c166a17ca4d7852e1ced8d53e9cb23f5593441274cae7ba041c Feb 15 20:33:42 crc kubenswrapper[4735]: I0215 20:33:42.629772 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.137584 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zrmz5" event={"ID":"35e018a0-8bf3-4725-a229-5a4196d4ae97","Type":"ContainerStarted","Data":"7ebf0a8fb4784c166a17ca4d7852e1ced8d53e9cb23f5593441274cae7ba041c"} Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.137750 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4n5d2" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.137879 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kjhzm" podUID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerName="registry-server" containerID="cri-o://b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96" gracePeriod=2 Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.204051 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-4n5d2"] Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.208895 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-4n5d2"] Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.240476 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.592171 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.697089 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-utilities\") pod \"db12b6bf-810b-44b6-b498-5550c7aaf23d\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.697259 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p54f\" (UniqueName: \"kubernetes.io/projected/db12b6bf-810b-44b6-b498-5550c7aaf23d-kube-api-access-4p54f\") pod \"db12b6bf-810b-44b6-b498-5550c7aaf23d\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.697345 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-catalog-content\") pod \"db12b6bf-810b-44b6-b498-5550c7aaf23d\" (UID: \"db12b6bf-810b-44b6-b498-5550c7aaf23d\") " Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.698343 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-utilities" (OuterVolumeSpecName: "utilities") pod "db12b6bf-810b-44b6-b498-5550c7aaf23d" (UID: "db12b6bf-810b-44b6-b498-5550c7aaf23d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.721146 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db12b6bf-810b-44b6-b498-5550c7aaf23d-kube-api-access-4p54f" (OuterVolumeSpecName: "kube-api-access-4p54f") pod "db12b6bf-810b-44b6-b498-5550c7aaf23d" (UID: "db12b6bf-810b-44b6-b498-5550c7aaf23d"). InnerVolumeSpecName "kube-api-access-4p54f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.758451 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db12b6bf-810b-44b6-b498-5550c7aaf23d" (UID: "db12b6bf-810b-44b6-b498-5550c7aaf23d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.800885 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.800918 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p54f\" (UniqueName: \"kubernetes.io/projected/db12b6bf-810b-44b6-b498-5550c7aaf23d-kube-api-access-4p54f\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.800930 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db12b6bf-810b-44b6-b498-5550c7aaf23d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.843309 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 15 20:33:43 crc kubenswrapper[4735]: I0215 20:33:43.843353 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.145662 4735 generic.go:334] "Generic (PLEG): container finished" podID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerID="b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96" exitCode=0 Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.146398 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjhzm" Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.147507 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjhzm" event={"ID":"db12b6bf-810b-44b6-b498-5550c7aaf23d","Type":"ContainerDied","Data":"b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96"} Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.147534 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjhzm" event={"ID":"db12b6bf-810b-44b6-b498-5550c7aaf23d","Type":"ContainerDied","Data":"f1fe1df17d6bf62067d0d48c2f63ab1d344b662db355a30978e08cc0c8245efb"} Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.147551 4735 scope.go:117] "RemoveContainer" containerID="b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96" Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.209594 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kjhzm"] Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.217343 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kjhzm"] Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.637184 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.752562 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.896503 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b063d1da-ed32-4c30-8b2a-2e65a4fcaceb" path="/var/lib/kubelet/pods/b063d1da-ed32-4c30-8b2a-2e65a4fcaceb/volumes" Feb 15 20:33:44 crc kubenswrapper[4735]: I0215 20:33:44.896886 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db12b6bf-810b-44b6-b498-5550c7aaf23d" path="/var/lib/kubelet/pods/db12b6bf-810b-44b6-b498-5550c7aaf23d/volumes" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.286819 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e7dc-account-create-update-mrg7s"] Feb 15 20:33:45 crc kubenswrapper[4735]: E0215 20:33:45.287184 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerName="extract-content" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.287196 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerName="extract-content" Feb 15 20:33:45 crc kubenswrapper[4735]: E0215 20:33:45.287213 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerName="extract-utilities" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.287219 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerName="extract-utilities" Feb 15 20:33:45 crc kubenswrapper[4735]: E0215 20:33:45.287257 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerName="registry-server" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.287263 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerName="registry-server" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.287403 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="db12b6bf-810b-44b6-b498-5550c7aaf23d" containerName="registry-server" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.288073 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.292484 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.298125 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e7dc-account-create-update-mrg7s"] Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.330412 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c606b7-ff0f-48d8-94c6-8beff123f81e-operator-scripts\") pod \"keystone-e7dc-account-create-update-mrg7s\" (UID: \"66c606b7-ff0f-48d8-94c6-8beff123f81e\") " pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.330521 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lppw\" (UniqueName: \"kubernetes.io/projected/66c606b7-ff0f-48d8-94c6-8beff123f81e-kube-api-access-8lppw\") pod \"keystone-e7dc-account-create-update-mrg7s\" (UID: \"66c606b7-ff0f-48d8-94c6-8beff123f81e\") " pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.332214 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-rtdvp"] Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.334164 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.345393 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-rtdvp"] Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.431801 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lppw\" (UniqueName: \"kubernetes.io/projected/66c606b7-ff0f-48d8-94c6-8beff123f81e-kube-api-access-8lppw\") pod \"keystone-e7dc-account-create-update-mrg7s\" (UID: \"66c606b7-ff0f-48d8-94c6-8beff123f81e\") " pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.431891 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c606b7-ff0f-48d8-94c6-8beff123f81e-operator-scripts\") pod \"keystone-e7dc-account-create-update-mrg7s\" (UID: \"66c606b7-ff0f-48d8-94c6-8beff123f81e\") " pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.431913 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpmcj\" (UniqueName: \"kubernetes.io/projected/49b95b5d-8023-49fd-9af3-828e64cbc1fc-kube-api-access-hpmcj\") pod \"keystone-db-create-rtdvp\" (UID: \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\") " pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.431940 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49b95b5d-8023-49fd-9af3-828e64cbc1fc-operator-scripts\") pod \"keystone-db-create-rtdvp\" (UID: \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\") " pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.432610 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c606b7-ff0f-48d8-94c6-8beff123f81e-operator-scripts\") pod \"keystone-e7dc-account-create-update-mrg7s\" (UID: \"66c606b7-ff0f-48d8-94c6-8beff123f81e\") " pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.471067 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lppw\" (UniqueName: \"kubernetes.io/projected/66c606b7-ff0f-48d8-94c6-8beff123f81e-kube-api-access-8lppw\") pod \"keystone-e7dc-account-create-update-mrg7s\" (UID: \"66c606b7-ff0f-48d8-94c6-8beff123f81e\") " pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.477034 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-9l2qw"] Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.478848 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.484533 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-9l2qw"] Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.533044 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49b95b5d-8023-49fd-9af3-828e64cbc1fc-operator-scripts\") pod \"keystone-db-create-rtdvp\" (UID: \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\") " pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.533088 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf282\" (UniqueName: \"kubernetes.io/projected/e4c17912-155f-4926-b2da-ee8bfe43a473-kube-api-access-qf282\") pod \"placement-db-create-9l2qw\" (UID: \"e4c17912-155f-4926-b2da-ee8bfe43a473\") " pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.533167 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4c17912-155f-4926-b2da-ee8bfe43a473-operator-scripts\") pod \"placement-db-create-9l2qw\" (UID: \"e4c17912-155f-4926-b2da-ee8bfe43a473\") " pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.533221 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpmcj\" (UniqueName: \"kubernetes.io/projected/49b95b5d-8023-49fd-9af3-828e64cbc1fc-kube-api-access-hpmcj\") pod \"keystone-db-create-rtdvp\" (UID: \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\") " pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.533820 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49b95b5d-8023-49fd-9af3-828e64cbc1fc-operator-scripts\") pod \"keystone-db-create-rtdvp\" (UID: \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\") " pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.561463 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpmcj\" (UniqueName: \"kubernetes.io/projected/49b95b5d-8023-49fd-9af3-828e64cbc1fc-kube-api-access-hpmcj\") pod \"keystone-db-create-rtdvp\" (UID: \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\") " pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.577413 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5b20-account-create-update-xbszc"] Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.578762 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.581396 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.592398 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5b20-account-create-update-xbszc"] Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.619909 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.635742 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71b4bdc2-22e8-4109-8210-782c0f222045-operator-scripts\") pod \"placement-5b20-account-create-update-xbszc\" (UID: \"71b4bdc2-22e8-4109-8210-782c0f222045\") " pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.635796 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf282\" (UniqueName: \"kubernetes.io/projected/e4c17912-155f-4926-b2da-ee8bfe43a473-kube-api-access-qf282\") pod \"placement-db-create-9l2qw\" (UID: \"e4c17912-155f-4926-b2da-ee8bfe43a473\") " pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.635955 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4c17912-155f-4926-b2da-ee8bfe43a473-operator-scripts\") pod \"placement-db-create-9l2qw\" (UID: \"e4c17912-155f-4926-b2da-ee8bfe43a473\") " pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.636004 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv2p5\" (UniqueName: \"kubernetes.io/projected/71b4bdc2-22e8-4109-8210-782c0f222045-kube-api-access-xv2p5\") pod \"placement-5b20-account-create-update-xbszc\" (UID: \"71b4bdc2-22e8-4109-8210-782c0f222045\") " pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.636554 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4c17912-155f-4926-b2da-ee8bfe43a473-operator-scripts\") pod \"placement-db-create-9l2qw\" (UID: \"e4c17912-155f-4926-b2da-ee8bfe43a473\") " pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.651427 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.652232 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf282\" (UniqueName: \"kubernetes.io/projected/e4c17912-155f-4926-b2da-ee8bfe43a473-kube-api-access-qf282\") pod \"placement-db-create-9l2qw\" (UID: \"e4c17912-155f-4926-b2da-ee8bfe43a473\") " pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.737853 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv2p5\" (UniqueName: \"kubernetes.io/projected/71b4bdc2-22e8-4109-8210-782c0f222045-kube-api-access-xv2p5\") pod \"placement-5b20-account-create-update-xbszc\" (UID: \"71b4bdc2-22e8-4109-8210-782c0f222045\") " pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.737922 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71b4bdc2-22e8-4109-8210-782c0f222045-operator-scripts\") pod \"placement-5b20-account-create-update-xbszc\" (UID: \"71b4bdc2-22e8-4109-8210-782c0f222045\") " pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.737981 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:45 crc kubenswrapper[4735]: E0215 20:33:45.738133 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 15 20:33:45 crc kubenswrapper[4735]: E0215 20:33:45.738145 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 15 20:33:45 crc kubenswrapper[4735]: E0215 20:33:45.738203 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift podName:a9642a7f-e8aa-4368-b2b9-3da6537c4d8f nodeName:}" failed. No retries permitted until 2026-02-15 20:33:53.738168332 +0000 UTC m=+1041.604183945 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift") pod "swift-storage-0" (UID: "a9642a7f-e8aa-4368-b2b9-3da6537c4d8f") : configmap "swift-ring-files" not found Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.740363 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71b4bdc2-22e8-4109-8210-782c0f222045-operator-scripts\") pod \"placement-5b20-account-create-update-xbszc\" (UID: \"71b4bdc2-22e8-4109-8210-782c0f222045\") " pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.754405 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv2p5\" (UniqueName: \"kubernetes.io/projected/71b4bdc2-22e8-4109-8210-782c0f222045-kube-api-access-xv2p5\") pod \"placement-5b20-account-create-update-xbszc\" (UID: \"71b4bdc2-22e8-4109-8210-782c0f222045\") " pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.813747 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:45 crc kubenswrapper[4735]: I0215 20:33:45.898200 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:46 crc kubenswrapper[4735]: I0215 20:33:46.358627 4735 scope.go:117] "RemoveContainer" containerID="c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b" Feb 15 20:33:46 crc kubenswrapper[4735]: I0215 20:33:46.472443 4735 scope.go:117] "RemoveContainer" containerID="6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8" Feb 15 20:33:46 crc kubenswrapper[4735]: I0215 20:33:46.655459 4735 scope.go:117] "RemoveContainer" containerID="b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96" Feb 15 20:33:46 crc kubenswrapper[4735]: E0215 20:33:46.657275 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96\": container with ID starting with b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96 not found: ID does not exist" containerID="b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96" Feb 15 20:33:46 crc kubenswrapper[4735]: I0215 20:33:46.657311 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96"} err="failed to get container status \"b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96\": rpc error: code = NotFound desc = could not find container \"b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96\": container with ID starting with b386d58d606d5e7070ab8cae318815beb91bbaa47b657f2e772018e7b54f7b96 not found: ID does not exist" Feb 15 20:33:46 crc kubenswrapper[4735]: I0215 20:33:46.657332 4735 scope.go:117] "RemoveContainer" containerID="c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b" Feb 15 20:33:46 crc kubenswrapper[4735]: E0215 20:33:46.660113 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b\": container with ID starting with c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b not found: ID does not exist" containerID="c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b" Feb 15 20:33:46 crc kubenswrapper[4735]: I0215 20:33:46.660144 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b"} err="failed to get container status \"c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b\": rpc error: code = NotFound desc = could not find container \"c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b\": container with ID starting with c589e805f86c3521255f41c5e522988bbe3a5745cd76e479baa42d7ba0889b0b not found: ID does not exist" Feb 15 20:33:46 crc kubenswrapper[4735]: I0215 20:33:46.660158 4735 scope.go:117] "RemoveContainer" containerID="6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8" Feb 15 20:33:46 crc kubenswrapper[4735]: E0215 20:33:46.660711 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8\": container with ID starting with 6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8 not found: ID does not exist" containerID="6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8" Feb 15 20:33:46 crc kubenswrapper[4735]: I0215 20:33:46.660752 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8"} err="failed to get container status \"6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8\": rpc error: code = NotFound desc = could not find container \"6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8\": container with ID starting with 6cae4f011f8fc1b6c035e40cca88e51c3e76dde5e44e6b9587c8264481f9a4f8 not found: ID does not exist" Feb 15 20:33:47 crc kubenswrapper[4735]: W0215 20:33:47.015521 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71b4bdc2_22e8_4109_8210_782c0f222045.slice/crio-2a69f050d4a2dd8ee6945617ba9d1913a799e72c6d2fb0bd2fbf5cbc2ac3ce1f WatchSource:0}: Error finding container 2a69f050d4a2dd8ee6945617ba9d1913a799e72c6d2fb0bd2fbf5cbc2ac3ce1f: Status 404 returned error can't find the container with id 2a69f050d4a2dd8ee6945617ba9d1913a799e72c6d2fb0bd2fbf5cbc2ac3ce1f Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.020583 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5b20-account-create-update-xbszc"] Feb 15 20:33:47 crc kubenswrapper[4735]: W0215 20:33:47.022151 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49b95b5d_8023_49fd_9af3_828e64cbc1fc.slice/crio-3c64d29d885272c38519ae4605a17a73e02a0c840a2a3c51a6fdd2646e346a99 WatchSource:0}: Error finding container 3c64d29d885272c38519ae4605a17a73e02a0c840a2a3c51a6fdd2646e346a99: Status 404 returned error can't find the container with id 3c64d29d885272c38519ae4605a17a73e02a0c840a2a3c51a6fdd2646e346a99 Feb 15 20:33:47 crc kubenswrapper[4735]: W0215 20:33:47.022445 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4c17912_155f_4926_b2da_ee8bfe43a473.slice/crio-4f393b0f44c69e67b59dc65773d5636f6011b880a7a2ed01ec1b55823500e71a WatchSource:0}: Error finding container 4f393b0f44c69e67b59dc65773d5636f6011b880a7a2ed01ec1b55823500e71a: Status 404 returned error can't find the container with id 4f393b0f44c69e67b59dc65773d5636f6011b880a7a2ed01ec1b55823500e71a Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.028295 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-rtdvp"] Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.031662 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-9l2qw"] Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.172134 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b20-account-create-update-xbszc" event={"ID":"71b4bdc2-22e8-4109-8210-782c0f222045","Type":"ContainerStarted","Data":"90ce3286cf31e407ce65270185e348d898a3df8f4f741d5c84b84148c2904fb3"} Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.172176 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b20-account-create-update-xbszc" event={"ID":"71b4bdc2-22e8-4109-8210-782c0f222045","Type":"ContainerStarted","Data":"2a69f050d4a2dd8ee6945617ba9d1913a799e72c6d2fb0bd2fbf5cbc2ac3ce1f"} Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.175391 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zrmz5" event={"ID":"35e018a0-8bf3-4725-a229-5a4196d4ae97","Type":"ContainerStarted","Data":"dfb793c89b775770fe14d0f1596d98cef43a59921eaf534990ad141fbe9a590e"} Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.179485 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9l2qw" event={"ID":"e4c17912-155f-4926-b2da-ee8bfe43a473","Type":"ContainerStarted","Data":"45da017dd2d568023786d1d5eb1eeff68778cb3d628305a3246665dd9d8bbb37"} Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.179523 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9l2qw" event={"ID":"e4c17912-155f-4926-b2da-ee8bfe43a473","Type":"ContainerStarted","Data":"4f393b0f44c69e67b59dc65773d5636f6011b880a7a2ed01ec1b55823500e71a"} Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.181465 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rtdvp" event={"ID":"49b95b5d-8023-49fd-9af3-828e64cbc1fc","Type":"ContainerStarted","Data":"a5b026cedf3008efb62514c2b8b28bcdba0a68446846e0d37243658973316b8b"} Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.181490 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rtdvp" event={"ID":"49b95b5d-8023-49fd-9af3-828e64cbc1fc","Type":"ContainerStarted","Data":"3c64d29d885272c38519ae4605a17a73e02a0c840a2a3c51a6fdd2646e346a99"} Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.210689 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5b20-account-create-update-xbszc" podStartSLOduration=2.210670941 podStartE2EDuration="2.210670941s" podCreationTimestamp="2026-02-15 20:33:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:33:47.197902809 +0000 UTC m=+1035.063918432" watchObservedRunningTime="2026-02-15 20:33:47.210670941 +0000 UTC m=+1035.076686574" Feb 15 20:33:47 crc kubenswrapper[4735]: W0215 20:33:47.220294 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66c606b7_ff0f_48d8_94c6_8beff123f81e.slice/crio-703cb41a837c0a4b67fa8a76581e2b401a51cfc8fa26268550051a3d6788449f WatchSource:0}: Error finding container 703cb41a837c0a4b67fa8a76581e2b401a51cfc8fa26268550051a3d6788449f: Status 404 returned error can't find the container with id 703cb41a837c0a4b67fa8a76581e2b401a51cfc8fa26268550051a3d6788449f Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.221162 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e7dc-account-create-update-mrg7s"] Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.228675 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-rtdvp" podStartSLOduration=2.228655344 podStartE2EDuration="2.228655344s" podCreationTimestamp="2026-02-15 20:33:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:33:47.207473205 +0000 UTC m=+1035.073488828" watchObservedRunningTime="2026-02-15 20:33:47.228655344 +0000 UTC m=+1035.094670967" Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.232994 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-9l2qw" podStartSLOduration=2.23298089 podStartE2EDuration="2.23298089s" podCreationTimestamp="2026-02-15 20:33:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:33:47.220641199 +0000 UTC m=+1035.086656822" watchObservedRunningTime="2026-02-15 20:33:47.23298089 +0000 UTC m=+1035.098996513" Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.246728 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-zrmz5" podStartSLOduration=2.310656818 podStartE2EDuration="6.246712378s" podCreationTimestamp="2026-02-15 20:33:41 +0000 UTC" firstStartedPulling="2026-02-15 20:33:42.554027397 +0000 UTC m=+1030.420043020" lastFinishedPulling="2026-02-15 20:33:46.490082907 +0000 UTC m=+1034.356098580" observedRunningTime="2026-02-15 20:33:47.242925327 +0000 UTC m=+1035.108940950" watchObservedRunningTime="2026-02-15 20:33:47.246712378 +0000 UTC m=+1035.112727991" Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.696178 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.909256 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:33:47 crc kubenswrapper[4735]: I0215 20:33:47.974292 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-62khd"] Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.194630 4735 generic.go:334] "Generic (PLEG): container finished" podID="66c606b7-ff0f-48d8-94c6-8beff123f81e" containerID="eb4d060cd69cd3d8740ffd482b2646fad6b20b0c95addca72d272e1cc473bf54" exitCode=0 Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.195088 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e7dc-account-create-update-mrg7s" event={"ID":"66c606b7-ff0f-48d8-94c6-8beff123f81e","Type":"ContainerDied","Data":"eb4d060cd69cd3d8740ffd482b2646fad6b20b0c95addca72d272e1cc473bf54"} Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.195120 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e7dc-account-create-update-mrg7s" event={"ID":"66c606b7-ff0f-48d8-94c6-8beff123f81e","Type":"ContainerStarted","Data":"703cb41a837c0a4b67fa8a76581e2b401a51cfc8fa26268550051a3d6788449f"} Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.197265 4735 generic.go:334] "Generic (PLEG): container finished" podID="e4c17912-155f-4926-b2da-ee8bfe43a473" containerID="45da017dd2d568023786d1d5eb1eeff68778cb3d628305a3246665dd9d8bbb37" exitCode=0 Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.197319 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9l2qw" event={"ID":"e4c17912-155f-4926-b2da-ee8bfe43a473","Type":"ContainerDied","Data":"45da017dd2d568023786d1d5eb1eeff68778cb3d628305a3246665dd9d8bbb37"} Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.205219 4735 generic.go:334] "Generic (PLEG): container finished" podID="49b95b5d-8023-49fd-9af3-828e64cbc1fc" containerID="a5b026cedf3008efb62514c2b8b28bcdba0a68446846e0d37243658973316b8b" exitCode=0 Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.205302 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rtdvp" event={"ID":"49b95b5d-8023-49fd-9af3-828e64cbc1fc","Type":"ContainerDied","Data":"a5b026cedf3008efb62514c2b8b28bcdba0a68446846e0d37243658973316b8b"} Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.238646 4735 generic.go:334] "Generic (PLEG): container finished" podID="71b4bdc2-22e8-4109-8210-782c0f222045" containerID="90ce3286cf31e407ce65270185e348d898a3df8f4f741d5c84b84148c2904fb3" exitCode=0 Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.243317 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b20-account-create-update-xbszc" event={"ID":"71b4bdc2-22e8-4109-8210-782c0f222045","Type":"ContainerDied","Data":"90ce3286cf31e407ce65270185e348d898a3df8f4f741d5c84b84148c2904fb3"} Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.243695 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d65f699f-62khd" podUID="2d4946ef-28a3-4831-adaf-ec2904168e44" containerName="dnsmasq-dns" containerID="cri-o://9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9" gracePeriod=10 Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.714487 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.809267 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4brwf\" (UniqueName: \"kubernetes.io/projected/2d4946ef-28a3-4831-adaf-ec2904168e44-kube-api-access-4brwf\") pod \"2d4946ef-28a3-4831-adaf-ec2904168e44\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.809323 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-dns-svc\") pod \"2d4946ef-28a3-4831-adaf-ec2904168e44\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.809507 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-config\") pod \"2d4946ef-28a3-4831-adaf-ec2904168e44\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.809552 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-ovsdbserver-nb\") pod \"2d4946ef-28a3-4831-adaf-ec2904168e44\" (UID: \"2d4946ef-28a3-4831-adaf-ec2904168e44\") " Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.819117 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d4946ef-28a3-4831-adaf-ec2904168e44-kube-api-access-4brwf" (OuterVolumeSpecName: "kube-api-access-4brwf") pod "2d4946ef-28a3-4831-adaf-ec2904168e44" (UID: "2d4946ef-28a3-4831-adaf-ec2904168e44"). InnerVolumeSpecName "kube-api-access-4brwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.866652 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-config" (OuterVolumeSpecName: "config") pod "2d4946ef-28a3-4831-adaf-ec2904168e44" (UID: "2d4946ef-28a3-4831-adaf-ec2904168e44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.867373 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2d4946ef-28a3-4831-adaf-ec2904168e44" (UID: "2d4946ef-28a3-4831-adaf-ec2904168e44"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.876180 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d4946ef-28a3-4831-adaf-ec2904168e44" (UID: "2d4946ef-28a3-4831-adaf-ec2904168e44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.926376 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4brwf\" (UniqueName: \"kubernetes.io/projected/2d4946ef-28a3-4831-adaf-ec2904168e44-kube-api-access-4brwf\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.926409 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.926419 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:48 crc kubenswrapper[4735]: I0215 20:33:48.926429 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d4946ef-28a3-4831-adaf-ec2904168e44-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.247376 4735 generic.go:334] "Generic (PLEG): container finished" podID="2d4946ef-28a3-4831-adaf-ec2904168e44" containerID="9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9" exitCode=0 Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.249529 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-62khd" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.251468 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-62khd" event={"ID":"2d4946ef-28a3-4831-adaf-ec2904168e44","Type":"ContainerDied","Data":"9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9"} Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.251515 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-62khd" event={"ID":"2d4946ef-28a3-4831-adaf-ec2904168e44","Type":"ContainerDied","Data":"3f49d81c4fffa5a4410671237df5491df071b81101e63347132aa689ab7f7fec"} Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.251544 4735 scope.go:117] "RemoveContainer" containerID="9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.286523 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-62khd"] Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.295798 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-62khd"] Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.298147 4735 scope.go:117] "RemoveContainer" containerID="8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.341705 4735 scope.go:117] "RemoveContainer" containerID="9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9" Feb 15 20:33:49 crc kubenswrapper[4735]: E0215 20:33:49.343134 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9\": container with ID starting with 9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9 not found: ID does not exist" containerID="9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.343162 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9"} err="failed to get container status \"9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9\": rpc error: code = NotFound desc = could not find container \"9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9\": container with ID starting with 9283a5d5b23dbd6630d521c1f62e32b270d5e1ab497ae89cd51d1dbe50edfef9 not found: ID does not exist" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.343186 4735 scope.go:117] "RemoveContainer" containerID="8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975" Feb 15 20:33:49 crc kubenswrapper[4735]: E0215 20:33:49.347424 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975\": container with ID starting with 8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975 not found: ID does not exist" containerID="8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.347453 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975"} err="failed to get container status \"8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975\": rpc error: code = NotFound desc = could not find container \"8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975\": container with ID starting with 8b33405915e6ee33c6518a0799ec1e9d327dcd9aee00ed685ef032c244b6b975 not found: ID does not exist" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.506177 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-whjxb"] Feb 15 20:33:49 crc kubenswrapper[4735]: E0215 20:33:49.506753 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4946ef-28a3-4831-adaf-ec2904168e44" containerName="init" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.506765 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4946ef-28a3-4831-adaf-ec2904168e44" containerName="init" Feb 15 20:33:49 crc kubenswrapper[4735]: E0215 20:33:49.506781 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4946ef-28a3-4831-adaf-ec2904168e44" containerName="dnsmasq-dns" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.506787 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4946ef-28a3-4831-adaf-ec2904168e44" containerName="dnsmasq-dns" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.506940 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4946ef-28a3-4831-adaf-ec2904168e44" containerName="dnsmasq-dns" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.507426 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-whjxb" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.541318 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-whjxb"] Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.600010 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.641993 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71b4bdc2-22e8-4109-8210-782c0f222045-operator-scripts\") pod \"71b4bdc2-22e8-4109-8210-782c0f222045\" (UID: \"71b4bdc2-22e8-4109-8210-782c0f222045\") " Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.642148 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv2p5\" (UniqueName: \"kubernetes.io/projected/71b4bdc2-22e8-4109-8210-782c0f222045-kube-api-access-xv2p5\") pod \"71b4bdc2-22e8-4109-8210-782c0f222045\" (UID: \"71b4bdc2-22e8-4109-8210-782c0f222045\") " Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.642437 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c860f14-d28c-4d02-8433-15c65ff6e981-operator-scripts\") pod \"glance-db-create-whjxb\" (UID: \"2c860f14-d28c-4d02-8433-15c65ff6e981\") " pod="openstack/glance-db-create-whjxb" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.642496 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9tlj\" (UniqueName: \"kubernetes.io/projected/2c860f14-d28c-4d02-8433-15c65ff6e981-kube-api-access-f9tlj\") pod \"glance-db-create-whjxb\" (UID: \"2c860f14-d28c-4d02-8433-15c65ff6e981\") " pod="openstack/glance-db-create-whjxb" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.643445 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71b4bdc2-22e8-4109-8210-782c0f222045-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "71b4bdc2-22e8-4109-8210-782c0f222045" (UID: "71b4bdc2-22e8-4109-8210-782c0f222045"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.644407 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-4a5c-account-create-update-tq6qj"] Feb 15 20:33:49 crc kubenswrapper[4735]: E0215 20:33:49.644781 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4bdc2-22e8-4109-8210-782c0f222045" containerName="mariadb-account-create-update" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.644793 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4bdc2-22e8-4109-8210-782c0f222045" containerName="mariadb-account-create-update" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.646006 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b4bdc2-22e8-4109-8210-782c0f222045-kube-api-access-xv2p5" (OuterVolumeSpecName: "kube-api-access-xv2p5") pod "71b4bdc2-22e8-4109-8210-782c0f222045" (UID: "71b4bdc2-22e8-4109-8210-782c0f222045"). InnerVolumeSpecName "kube-api-access-xv2p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.652898 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4bdc2-22e8-4109-8210-782c0f222045" containerName="mariadb-account-create-update" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.653482 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.661243 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.693040 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4a5c-account-create-update-tq6qj"] Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.744125 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-operator-scripts\") pod \"glance-4a5c-account-create-update-tq6qj\" (UID: \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\") " pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.744177 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9tlj\" (UniqueName: \"kubernetes.io/projected/2c860f14-d28c-4d02-8433-15c65ff6e981-kube-api-access-f9tlj\") pod \"glance-db-create-whjxb\" (UID: \"2c860f14-d28c-4d02-8433-15c65ff6e981\") " pod="openstack/glance-db-create-whjxb" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.744365 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9szk7\" (UniqueName: \"kubernetes.io/projected/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-kube-api-access-9szk7\") pod \"glance-4a5c-account-create-update-tq6qj\" (UID: \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\") " pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.744454 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c860f14-d28c-4d02-8433-15c65ff6e981-operator-scripts\") pod \"glance-db-create-whjxb\" (UID: \"2c860f14-d28c-4d02-8433-15c65ff6e981\") " pod="openstack/glance-db-create-whjxb" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.745151 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv2p5\" (UniqueName: \"kubernetes.io/projected/71b4bdc2-22e8-4109-8210-782c0f222045-kube-api-access-xv2p5\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.745610 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71b4bdc2-22e8-4109-8210-782c0f222045-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.748344 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c860f14-d28c-4d02-8433-15c65ff6e981-operator-scripts\") pod \"glance-db-create-whjxb\" (UID: \"2c860f14-d28c-4d02-8433-15c65ff6e981\") " pod="openstack/glance-db-create-whjxb" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.764093 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9tlj\" (UniqueName: \"kubernetes.io/projected/2c860f14-d28c-4d02-8433-15c65ff6e981-kube-api-access-f9tlj\") pod \"glance-db-create-whjxb\" (UID: \"2c860f14-d28c-4d02-8433-15c65ff6e981\") " pod="openstack/glance-db-create-whjxb" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.841732 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.847036 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9szk7\" (UniqueName: \"kubernetes.io/projected/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-kube-api-access-9szk7\") pod \"glance-4a5c-account-create-update-tq6qj\" (UID: \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\") " pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.847171 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-operator-scripts\") pod \"glance-4a5c-account-create-update-tq6qj\" (UID: \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\") " pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.848718 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-operator-scripts\") pod \"glance-4a5c-account-create-update-tq6qj\" (UID: \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\") " pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.852685 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-whjxb" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.855601 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.872925 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.877636 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9szk7\" (UniqueName: \"kubernetes.io/projected/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-kube-api-access-9szk7\") pod \"glance-4a5c-account-create-update-tq6qj\" (UID: \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\") " pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.948073 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpmcj\" (UniqueName: \"kubernetes.io/projected/49b95b5d-8023-49fd-9af3-828e64cbc1fc-kube-api-access-hpmcj\") pod \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\" (UID: \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\") " Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.948123 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c606b7-ff0f-48d8-94c6-8beff123f81e-operator-scripts\") pod \"66c606b7-ff0f-48d8-94c6-8beff123f81e\" (UID: \"66c606b7-ff0f-48d8-94c6-8beff123f81e\") " Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.948211 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4c17912-155f-4926-b2da-ee8bfe43a473-operator-scripts\") pod \"e4c17912-155f-4926-b2da-ee8bfe43a473\" (UID: \"e4c17912-155f-4926-b2da-ee8bfe43a473\") " Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.948322 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49b95b5d-8023-49fd-9af3-828e64cbc1fc-operator-scripts\") pod \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\" (UID: \"49b95b5d-8023-49fd-9af3-828e64cbc1fc\") " Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.948403 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf282\" (UniqueName: \"kubernetes.io/projected/e4c17912-155f-4926-b2da-ee8bfe43a473-kube-api-access-qf282\") pod \"e4c17912-155f-4926-b2da-ee8bfe43a473\" (UID: \"e4c17912-155f-4926-b2da-ee8bfe43a473\") " Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.948441 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lppw\" (UniqueName: \"kubernetes.io/projected/66c606b7-ff0f-48d8-94c6-8beff123f81e-kube-api-access-8lppw\") pod \"66c606b7-ff0f-48d8-94c6-8beff123f81e\" (UID: \"66c606b7-ff0f-48d8-94c6-8beff123f81e\") " Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.950597 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c606b7-ff0f-48d8-94c6-8beff123f81e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "66c606b7-ff0f-48d8-94c6-8beff123f81e" (UID: "66c606b7-ff0f-48d8-94c6-8beff123f81e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.950614 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49b95b5d-8023-49fd-9af3-828e64cbc1fc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "49b95b5d-8023-49fd-9af3-828e64cbc1fc" (UID: "49b95b5d-8023-49fd-9af3-828e64cbc1fc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.950936 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4c17912-155f-4926-b2da-ee8bfe43a473-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e4c17912-155f-4926-b2da-ee8bfe43a473" (UID: "e4c17912-155f-4926-b2da-ee8bfe43a473"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.954678 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b95b5d-8023-49fd-9af3-828e64cbc1fc-kube-api-access-hpmcj" (OuterVolumeSpecName: "kube-api-access-hpmcj") pod "49b95b5d-8023-49fd-9af3-828e64cbc1fc" (UID: "49b95b5d-8023-49fd-9af3-828e64cbc1fc"). InnerVolumeSpecName "kube-api-access-hpmcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.954713 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66c606b7-ff0f-48d8-94c6-8beff123f81e-kube-api-access-8lppw" (OuterVolumeSpecName: "kube-api-access-8lppw") pod "66c606b7-ff0f-48d8-94c6-8beff123f81e" (UID: "66c606b7-ff0f-48d8-94c6-8beff123f81e"). InnerVolumeSpecName "kube-api-access-8lppw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:49 crc kubenswrapper[4735]: I0215 20:33:49.957119 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4c17912-155f-4926-b2da-ee8bfe43a473-kube-api-access-qf282" (OuterVolumeSpecName: "kube-api-access-qf282") pod "e4c17912-155f-4926-b2da-ee8bfe43a473" (UID: "e4c17912-155f-4926-b2da-ee8bfe43a473"). InnerVolumeSpecName "kube-api-access-qf282". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.050177 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.051922 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpmcj\" (UniqueName: \"kubernetes.io/projected/49b95b5d-8023-49fd-9af3-828e64cbc1fc-kube-api-access-hpmcj\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.051956 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c606b7-ff0f-48d8-94c6-8beff123f81e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.051972 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4c17912-155f-4926-b2da-ee8bfe43a473-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.051982 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49b95b5d-8023-49fd-9af3-828e64cbc1fc-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.052002 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf282\" (UniqueName: \"kubernetes.io/projected/e4c17912-155f-4926-b2da-ee8bfe43a473-kube-api-access-qf282\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.052013 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lppw\" (UniqueName: \"kubernetes.io/projected/66c606b7-ff0f-48d8-94c6-8beff123f81e-kube-api-access-8lppw\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.259209 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9l2qw" event={"ID":"e4c17912-155f-4926-b2da-ee8bfe43a473","Type":"ContainerDied","Data":"4f393b0f44c69e67b59dc65773d5636f6011b880a7a2ed01ec1b55823500e71a"} Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.259249 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f393b0f44c69e67b59dc65773d5636f6011b880a7a2ed01ec1b55823500e71a" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.259299 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9l2qw" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.260698 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rtdvp" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.260736 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rtdvp" event={"ID":"49b95b5d-8023-49fd-9af3-828e64cbc1fc","Type":"ContainerDied","Data":"3c64d29d885272c38519ae4605a17a73e02a0c840a2a3c51a6fdd2646e346a99"} Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.260776 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c64d29d885272c38519ae4605a17a73e02a0c840a2a3c51a6fdd2646e346a99" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.261911 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b20-account-create-update-xbszc" event={"ID":"71b4bdc2-22e8-4109-8210-782c0f222045","Type":"ContainerDied","Data":"2a69f050d4a2dd8ee6945617ba9d1913a799e72c6d2fb0bd2fbf5cbc2ac3ce1f"} Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.261936 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a69f050d4a2dd8ee6945617ba9d1913a799e72c6d2fb0bd2fbf5cbc2ac3ce1f" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.262005 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5b20-account-create-update-xbszc" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.272178 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e7dc-account-create-update-mrg7s" event={"ID":"66c606b7-ff0f-48d8-94c6-8beff123f81e","Type":"ContainerDied","Data":"703cb41a837c0a4b67fa8a76581e2b401a51cfc8fa26268550051a3d6788449f"} Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.272225 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="703cb41a837c0a4b67fa8a76581e2b401a51cfc8fa26268550051a3d6788449f" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.272308 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e7dc-account-create-update-mrg7s" Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.333444 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-whjxb"] Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.494231 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4a5c-account-create-update-tq6qj"] Feb 15 20:33:50 crc kubenswrapper[4735]: I0215 20:33:50.898776 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d4946ef-28a3-4831-adaf-ec2904168e44" path="/var/lib/kubelet/pods/2d4946ef-28a3-4831-adaf-ec2904168e44/volumes" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.158893 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-dsp68"] Feb 15 20:33:51 crc kubenswrapper[4735]: E0215 20:33:51.159922 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b95b5d-8023-49fd-9af3-828e64cbc1fc" containerName="mariadb-database-create" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.160065 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b95b5d-8023-49fd-9af3-828e64cbc1fc" containerName="mariadb-database-create" Feb 15 20:33:51 crc kubenswrapper[4735]: E0215 20:33:51.160178 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4c17912-155f-4926-b2da-ee8bfe43a473" containerName="mariadb-database-create" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.160260 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4c17912-155f-4926-b2da-ee8bfe43a473" containerName="mariadb-database-create" Feb 15 20:33:51 crc kubenswrapper[4735]: E0215 20:33:51.160361 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c606b7-ff0f-48d8-94c6-8beff123f81e" containerName="mariadb-account-create-update" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.160558 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c606b7-ff0f-48d8-94c6-8beff123f81e" containerName="mariadb-account-create-update" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.161934 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b95b5d-8023-49fd-9af3-828e64cbc1fc" containerName="mariadb-database-create" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.162084 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4c17912-155f-4926-b2da-ee8bfe43a473" containerName="mariadb-database-create" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.162181 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="66c606b7-ff0f-48d8-94c6-8beff123f81e" containerName="mariadb-account-create-update" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.162826 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.166935 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.179900 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-dsp68"] Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.274081 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-operator-scripts\") pod \"root-account-create-update-dsp68\" (UID: \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\") " pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.274178 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8msf\" (UniqueName: \"kubernetes.io/projected/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-kube-api-access-c8msf\") pod \"root-account-create-update-dsp68\" (UID: \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\") " pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.281221 4735 generic.go:334] "Generic (PLEG): container finished" podID="2c860f14-d28c-4d02-8433-15c65ff6e981" containerID="eb389e6c8ab356ba85b72553b28fdd28b0db1572c97b36a3dc3fb08ccaa72bc1" exitCode=0 Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.281321 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-whjxb" event={"ID":"2c860f14-d28c-4d02-8433-15c65ff6e981","Type":"ContainerDied","Data":"eb389e6c8ab356ba85b72553b28fdd28b0db1572c97b36a3dc3fb08ccaa72bc1"} Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.281352 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-whjxb" event={"ID":"2c860f14-d28c-4d02-8433-15c65ff6e981","Type":"ContainerStarted","Data":"a06aeb5d7087a36dc79beb1a8269f19a237e264200843f6bfadedbce6a625349"} Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.283387 4735 generic.go:334] "Generic (PLEG): container finished" podID="9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9" containerID="045ff36bcd7e2a5303e6b80261e9ada2e1904c716782a5889b1317f9b7b7e67f" exitCode=0 Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.283590 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4a5c-account-create-update-tq6qj" event={"ID":"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9","Type":"ContainerDied","Data":"045ff36bcd7e2a5303e6b80261e9ada2e1904c716782a5889b1317f9b7b7e67f"} Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.283746 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4a5c-account-create-update-tq6qj" event={"ID":"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9","Type":"ContainerStarted","Data":"79a29bdcbb98ce12955be3f5c4d8b724057303982609b0c2e9c73a4aaae92c1b"} Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.376097 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-operator-scripts\") pod \"root-account-create-update-dsp68\" (UID: \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\") " pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.376174 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8msf\" (UniqueName: \"kubernetes.io/projected/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-kube-api-access-c8msf\") pod \"root-account-create-update-dsp68\" (UID: \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\") " pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.377904 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-operator-scripts\") pod \"root-account-create-update-dsp68\" (UID: \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\") " pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.410684 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8msf\" (UniqueName: \"kubernetes.io/projected/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-kube-api-access-c8msf\") pod \"root-account-create-update-dsp68\" (UID: \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\") " pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.495083 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:51 crc kubenswrapper[4735]: I0215 20:33:51.958187 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-dsp68"] Feb 15 20:33:51 crc kubenswrapper[4735]: W0215 20:33:51.981172 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ad0eeb8_c930_4cb3_a2f4_37e6934618ce.slice/crio-ac85ccc05712cff5ffece727375cc5d471aed42579832f60b72cab605d62ca68 WatchSource:0}: Error finding container ac85ccc05712cff5ffece727375cc5d471aed42579832f60b72cab605d62ca68: Status 404 returned error can't find the container with id ac85ccc05712cff5ffece727375cc5d471aed42579832f60b72cab605d62ca68 Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.294170 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dsp68" event={"ID":"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce","Type":"ContainerStarted","Data":"279b8c3fba818d33c3c3e9ea5ef7e22c36679f2add8fd9f8c9f2d3aae9857205"} Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.294466 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dsp68" event={"ID":"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce","Type":"ContainerStarted","Data":"ac85ccc05712cff5ffece727375cc5d471aed42579832f60b72cab605d62ca68"} Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.313642 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-dsp68" podStartSLOduration=1.313618914 podStartE2EDuration="1.313618914s" podCreationTimestamp="2026-02-15 20:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:33:52.309013859 +0000 UTC m=+1040.175029492" watchObservedRunningTime="2026-02-15 20:33:52.313618914 +0000 UTC m=+1040.179634547" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.665975 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.670680 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-whjxb" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.718758 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c860f14-d28c-4d02-8433-15c65ff6e981-operator-scripts\") pod \"2c860f14-d28c-4d02-8433-15c65ff6e981\" (UID: \"2c860f14-d28c-4d02-8433-15c65ff6e981\") " Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.718846 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9szk7\" (UniqueName: \"kubernetes.io/projected/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-kube-api-access-9szk7\") pod \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\" (UID: \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\") " Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.718978 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-operator-scripts\") pod \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\" (UID: \"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9\") " Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.719058 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9tlj\" (UniqueName: \"kubernetes.io/projected/2c860f14-d28c-4d02-8433-15c65ff6e981-kube-api-access-f9tlj\") pod \"2c860f14-d28c-4d02-8433-15c65ff6e981\" (UID: \"2c860f14-d28c-4d02-8433-15c65ff6e981\") " Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.719496 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9" (UID: "9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.720127 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c860f14-d28c-4d02-8433-15c65ff6e981-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2c860f14-d28c-4d02-8433-15c65ff6e981" (UID: "2c860f14-d28c-4d02-8433-15c65ff6e981"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.728226 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-kube-api-access-9szk7" (OuterVolumeSpecName: "kube-api-access-9szk7") pod "9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9" (UID: "9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9"). InnerVolumeSpecName "kube-api-access-9szk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.732301 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c860f14-d28c-4d02-8433-15c65ff6e981-kube-api-access-f9tlj" (OuterVolumeSpecName: "kube-api-access-f9tlj") pod "2c860f14-d28c-4d02-8433-15c65ff6e981" (UID: "2c860f14-d28c-4d02-8433-15c65ff6e981"). InnerVolumeSpecName "kube-api-access-f9tlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.821163 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9szk7\" (UniqueName: \"kubernetes.io/projected/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-kube-api-access-9szk7\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.821192 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.821202 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9tlj\" (UniqueName: \"kubernetes.io/projected/2c860f14-d28c-4d02-8433-15c65ff6e981-kube-api-access-f9tlj\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:52 crc kubenswrapper[4735]: I0215 20:33:52.821210 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c860f14-d28c-4d02-8433-15c65ff6e981-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:53 crc kubenswrapper[4735]: I0215 20:33:53.301522 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-whjxb" event={"ID":"2c860f14-d28c-4d02-8433-15c65ff6e981","Type":"ContainerDied","Data":"a06aeb5d7087a36dc79beb1a8269f19a237e264200843f6bfadedbce6a625349"} Feb 15 20:33:53 crc kubenswrapper[4735]: I0215 20:33:53.301977 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a06aeb5d7087a36dc79beb1a8269f19a237e264200843f6bfadedbce6a625349" Feb 15 20:33:53 crc kubenswrapper[4735]: I0215 20:33:53.301550 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-whjxb" Feb 15 20:33:53 crc kubenswrapper[4735]: I0215 20:33:53.304452 4735 generic.go:334] "Generic (PLEG): container finished" podID="9ad0eeb8-c930-4cb3-a2f4-37e6934618ce" containerID="279b8c3fba818d33c3c3e9ea5ef7e22c36679f2add8fd9f8c9f2d3aae9857205" exitCode=0 Feb 15 20:33:53 crc kubenswrapper[4735]: I0215 20:33:53.304518 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dsp68" event={"ID":"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce","Type":"ContainerDied","Data":"279b8c3fba818d33c3c3e9ea5ef7e22c36679f2add8fd9f8c9f2d3aae9857205"} Feb 15 20:33:53 crc kubenswrapper[4735]: I0215 20:33:53.306361 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4a5c-account-create-update-tq6qj" event={"ID":"9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9","Type":"ContainerDied","Data":"79a29bdcbb98ce12955be3f5c4d8b724057303982609b0c2e9c73a4aaae92c1b"} Feb 15 20:33:53 crc kubenswrapper[4735]: I0215 20:33:53.306469 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79a29bdcbb98ce12955be3f5c4d8b724057303982609b0c2e9c73a4aaae92c1b" Feb 15 20:33:53 crc kubenswrapper[4735]: I0215 20:33:53.306481 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4a5c-account-create-update-tq6qj" Feb 15 20:33:53 crc kubenswrapper[4735]: I0215 20:33:53.837119 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:33:53 crc kubenswrapper[4735]: E0215 20:33:53.837312 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 15 20:33:53 crc kubenswrapper[4735]: E0215 20:33:53.837355 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 15 20:33:53 crc kubenswrapper[4735]: E0215 20:33:53.837441 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift podName:a9642a7f-e8aa-4368-b2b9-3da6537c4d8f nodeName:}" failed. No retries permitted until 2026-02-15 20:34:09.837419052 +0000 UTC m=+1057.703434695 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift") pod "swift-storage-0" (UID: "a9642a7f-e8aa-4368-b2b9-3da6537c4d8f") : configmap "swift-ring-files" not found Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.319769 4735 generic.go:334] "Generic (PLEG): container finished" podID="35e018a0-8bf3-4725-a229-5a4196d4ae97" containerID="dfb793c89b775770fe14d0f1596d98cef43a59921eaf534990ad141fbe9a590e" exitCode=0 Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.319877 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zrmz5" event={"ID":"35e018a0-8bf3-4725-a229-5a4196d4ae97","Type":"ContainerDied","Data":"dfb793c89b775770fe14d0f1596d98cef43a59921eaf534990ad141fbe9a590e"} Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.770370 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.860512 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8msf\" (UniqueName: \"kubernetes.io/projected/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-kube-api-access-c8msf\") pod \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\" (UID: \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\") " Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.860583 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-operator-scripts\") pod \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\" (UID: \"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce\") " Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.861410 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9ad0eeb8-c930-4cb3-a2f4-37e6934618ce" (UID: "9ad0eeb8-c930-4cb3-a2f4-37e6934618ce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.867076 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-djbjm"] Feb 15 20:33:54 crc kubenswrapper[4735]: E0215 20:33:54.867436 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9" containerName="mariadb-account-create-update" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.867456 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9" containerName="mariadb-account-create-update" Feb 15 20:33:54 crc kubenswrapper[4735]: E0215 20:33:54.867476 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c860f14-d28c-4d02-8433-15c65ff6e981" containerName="mariadb-database-create" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.867485 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c860f14-d28c-4d02-8433-15c65ff6e981" containerName="mariadb-database-create" Feb 15 20:33:54 crc kubenswrapper[4735]: E0215 20:33:54.867502 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad0eeb8-c930-4cb3-a2f4-37e6934618ce" containerName="mariadb-account-create-update" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.867510 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad0eeb8-c930-4cb3-a2f4-37e6934618ce" containerName="mariadb-account-create-update" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.867722 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ad0eeb8-c930-4cb3-a2f4-37e6934618ce" containerName="mariadb-account-create-update" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.867744 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9" containerName="mariadb-account-create-update" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.867762 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c860f14-d28c-4d02-8433-15c65ff6e981" containerName="mariadb-database-create" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.868381 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.872046 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.875353 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kfjgg" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.879067 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-kube-api-access-c8msf" (OuterVolumeSpecName: "kube-api-access-c8msf") pod "9ad0eeb8-c930-4cb3-a2f4-37e6934618ce" (UID: "9ad0eeb8-c930-4cb3-a2f4-37e6934618ce"). InnerVolumeSpecName "kube-api-access-c8msf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.880911 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-djbjm"] Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.961925 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-config-data\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.961991 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnqnl\" (UniqueName: \"kubernetes.io/projected/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-kube-api-access-fnqnl\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.962083 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-db-sync-config-data\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.963380 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-combined-ca-bundle\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.963979 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8msf\" (UniqueName: \"kubernetes.io/projected/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-kube-api-access-c8msf\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:54 crc kubenswrapper[4735]: I0215 20:33:54.964021 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.065699 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-config-data\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.065752 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnqnl\" (UniqueName: \"kubernetes.io/projected/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-kube-api-access-fnqnl\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.065810 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-db-sync-config-data\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.065845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-combined-ca-bundle\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.070430 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-config-data\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.070426 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-combined-ca-bundle\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.070855 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-db-sync-config-data\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.088184 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnqnl\" (UniqueName: \"kubernetes.io/projected/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-kube-api-access-fnqnl\") pod \"glance-db-sync-djbjm\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.218805 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-djbjm" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.334272 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dsp68" event={"ID":"9ad0eeb8-c930-4cb3-a2f4-37e6934618ce","Type":"ContainerDied","Data":"ac85ccc05712cff5ffece727375cc5d471aed42579832f60b72cab605d62ca68"} Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.334571 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac85ccc05712cff5ffece727375cc5d471aed42579832f60b72cab605d62ca68" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.334312 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dsp68" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.590739 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-djbjm"] Feb 15 20:33:55 crc kubenswrapper[4735]: W0215 20:33:55.595192 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe509f30_9a8a_4cd2_8bf7_b8132bd92a97.slice/crio-2a9d1850d4c5ac17736735464e117a308cd9d9c276819788c57eaed404806951 WatchSource:0}: Error finding container 2a9d1850d4c5ac17736735464e117a308cd9d9c276819788c57eaed404806951: Status 404 returned error can't find the container with id 2a9d1850d4c5ac17736735464e117a308cd9d9c276819788c57eaed404806951 Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.608367 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.675429 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-dispersionconf\") pod \"35e018a0-8bf3-4725-a229-5a4196d4ae97\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.675692 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-swiftconf\") pod \"35e018a0-8bf3-4725-a229-5a4196d4ae97\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.675728 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jdrb\" (UniqueName: \"kubernetes.io/projected/35e018a0-8bf3-4725-a229-5a4196d4ae97-kube-api-access-8jdrb\") pod \"35e018a0-8bf3-4725-a229-5a4196d4ae97\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.675755 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35e018a0-8bf3-4725-a229-5a4196d4ae97-etc-swift\") pod \"35e018a0-8bf3-4725-a229-5a4196d4ae97\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.676617 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e018a0-8bf3-4725-a229-5a4196d4ae97-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "35e018a0-8bf3-4725-a229-5a4196d4ae97" (UID: "35e018a0-8bf3-4725-a229-5a4196d4ae97"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.676658 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-scripts\") pod \"35e018a0-8bf3-4725-a229-5a4196d4ae97\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.676699 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-ring-data-devices\") pod \"35e018a0-8bf3-4725-a229-5a4196d4ae97\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.677830 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "35e018a0-8bf3-4725-a229-5a4196d4ae97" (UID: "35e018a0-8bf3-4725-a229-5a4196d4ae97"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.677922 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-combined-ca-bundle\") pod \"35e018a0-8bf3-4725-a229-5a4196d4ae97\" (UID: \"35e018a0-8bf3-4725-a229-5a4196d4ae97\") " Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.678526 4735 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35e018a0-8bf3-4725-a229-5a4196d4ae97-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.678539 4735 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.682275 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e018a0-8bf3-4725-a229-5a4196d4ae97-kube-api-access-8jdrb" (OuterVolumeSpecName: "kube-api-access-8jdrb") pod "35e018a0-8bf3-4725-a229-5a4196d4ae97" (UID: "35e018a0-8bf3-4725-a229-5a4196d4ae97"). InnerVolumeSpecName "kube-api-access-8jdrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.684108 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "35e018a0-8bf3-4725-a229-5a4196d4ae97" (UID: "35e018a0-8bf3-4725-a229-5a4196d4ae97"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.699557 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-scripts" (OuterVolumeSpecName: "scripts") pod "35e018a0-8bf3-4725-a229-5a4196d4ae97" (UID: "35e018a0-8bf3-4725-a229-5a4196d4ae97"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.709173 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35e018a0-8bf3-4725-a229-5a4196d4ae97" (UID: "35e018a0-8bf3-4725-a229-5a4196d4ae97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.730026 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "35e018a0-8bf3-4725-a229-5a4196d4ae97" (UID: "35e018a0-8bf3-4725-a229-5a4196d4ae97"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.780575 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.780858 4735 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.780886 4735 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35e018a0-8bf3-4725-a229-5a4196d4ae97-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.780905 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jdrb\" (UniqueName: \"kubernetes.io/projected/35e018a0-8bf3-4725-a229-5a4196d4ae97-kube-api-access-8jdrb\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:55 crc kubenswrapper[4735]: I0215 20:33:55.780926 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35e018a0-8bf3-4725-a229-5a4196d4ae97-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:33:56 crc kubenswrapper[4735]: I0215 20:33:56.343564 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zrmz5" event={"ID":"35e018a0-8bf3-4725-a229-5a4196d4ae97","Type":"ContainerDied","Data":"7ebf0a8fb4784c166a17ca4d7852e1ced8d53e9cb23f5593441274cae7ba041c"} Feb 15 20:33:56 crc kubenswrapper[4735]: I0215 20:33:56.343596 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ebf0a8fb4784c166a17ca4d7852e1ced8d53e9cb23f5593441274cae7ba041c" Feb 15 20:33:56 crc kubenswrapper[4735]: I0215 20:33:56.343648 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zrmz5" Feb 15 20:33:56 crc kubenswrapper[4735]: I0215 20:33:56.348698 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-djbjm" event={"ID":"be509f30-9a8a-4cd2-8bf7-b8132bd92a97","Type":"ContainerStarted","Data":"2a9d1850d4c5ac17736735464e117a308cd9d9c276819788c57eaed404806951"} Feb 15 20:33:57 crc kubenswrapper[4735]: I0215 20:33:57.473878 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-dsp68"] Feb 15 20:33:57 crc kubenswrapper[4735]: I0215 20:33:57.480216 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-dsp68"] Feb 15 20:33:57 crc kubenswrapper[4735]: I0215 20:33:57.969985 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 15 20:33:58 crc kubenswrapper[4735]: I0215 20:33:58.897859 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ad0eeb8-c930-4cb3-a2f4-37e6934618ce" path="/var/lib/kubelet/pods/9ad0eeb8-c930-4cb3-a2f4-37e6934618ce/volumes" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.479934 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-frck7"] Feb 15 20:34:02 crc kubenswrapper[4735]: E0215 20:34:02.480660 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e018a0-8bf3-4725-a229-5a4196d4ae97" containerName="swift-ring-rebalance" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.480673 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e018a0-8bf3-4725-a229-5a4196d4ae97" containerName="swift-ring-rebalance" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.480817 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e018a0-8bf3-4725-a229-5a4196d4ae97" containerName="swift-ring-rebalance" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.481286 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-frck7" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.486288 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.496135 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-frck7"] Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.500438 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-operator-scripts\") pod \"root-account-create-update-frck7\" (UID: \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\") " pod="openstack/root-account-create-update-frck7" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.500483 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc6dn\" (UniqueName: \"kubernetes.io/projected/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-kube-api-access-rc6dn\") pod \"root-account-create-update-frck7\" (UID: \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\") " pod="openstack/root-account-create-update-frck7" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.601722 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-operator-scripts\") pod \"root-account-create-update-frck7\" (UID: \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\") " pod="openstack/root-account-create-update-frck7" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.601767 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc6dn\" (UniqueName: \"kubernetes.io/projected/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-kube-api-access-rc6dn\") pod \"root-account-create-update-frck7\" (UID: \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\") " pod="openstack/root-account-create-update-frck7" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.602460 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-operator-scripts\") pod \"root-account-create-update-frck7\" (UID: \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\") " pod="openstack/root-account-create-update-frck7" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.620531 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc6dn\" (UniqueName: \"kubernetes.io/projected/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-kube-api-access-rc6dn\") pod \"root-account-create-update-frck7\" (UID: \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\") " pod="openstack/root-account-create-update-frck7" Feb 15 20:34:02 crc kubenswrapper[4735]: I0215 20:34:02.809499 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-frck7" Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.438555 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c7b91a3-10d6-465d-aff4-4855d697b54f" containerID="528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca" exitCode=0 Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.438838 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5c7b91a3-10d6-465d-aff4-4855d697b54f","Type":"ContainerDied","Data":"528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca"} Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.442252 4735 generic.go:334] "Generic (PLEG): container finished" podID="dc3c285a-910a-4cb1-968e-636216557d09" containerID="6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a" exitCode=0 Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.442297 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc3c285a-910a-4cb1-968e-636216557d09","Type":"ContainerDied","Data":"6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a"} Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.536364 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-tk256" podUID="df76a192-fa00-4146-8f3a-8f7fa6798fec" containerName="ovn-controller" probeResult="failure" output=< Feb 15 20:34:04 crc kubenswrapper[4735]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 15 20:34:04 crc kubenswrapper[4735]: > Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.592258 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.602043 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fgcfc" Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.841430 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tk256-config-8wrkv"] Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.842385 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.846478 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 15 20:34:04 crc kubenswrapper[4735]: I0215 20:34:04.862240 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tk256-config-8wrkv"] Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.041253 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5lc2\" (UniqueName: \"kubernetes.io/projected/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-kube-api-access-p5lc2\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.041296 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-log-ovn\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.041331 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run-ovn\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.041675 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-scripts\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.041874 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-additional-scripts\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.041916 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.143253 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5lc2\" (UniqueName: \"kubernetes.io/projected/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-kube-api-access-p5lc2\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.143316 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-log-ovn\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.143376 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run-ovn\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.143588 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-scripts\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.143649 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-additional-scripts\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.143674 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.144048 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run-ovn\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.144791 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-additional-scripts\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.159902 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-scripts\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.160124 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-log-ovn\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.160160 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.166409 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5lc2\" (UniqueName: \"kubernetes.io/projected/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-kube-api-access-p5lc2\") pod \"ovn-controller-tk256-config-8wrkv\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:05 crc kubenswrapper[4735]: I0215 20:34:05.463250 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.039300 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-frck7"] Feb 15 20:34:08 crc kubenswrapper[4735]: W0215 20:34:08.040471 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8874208_49c3_4e6b_bdb3_ef40de3d9ec7.slice/crio-cc959deef23ce3450971377c27606f77bcd82533e21f1606035f874f18fdb408 WatchSource:0}: Error finding container cc959deef23ce3450971377c27606f77bcd82533e21f1606035f874f18fdb408: Status 404 returned error can't find the container with id cc959deef23ce3450971377c27606f77bcd82533e21f1606035f874f18fdb408 Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.133584 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tk256-config-8wrkv"] Feb 15 20:34:08 crc kubenswrapper[4735]: W0215 20:34:08.135440 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1bbe9fa_32f5_4a8e_8ba8_39f88fffbc68.slice/crio-67749dc1b4bb3dfdc39c890d1e54352654f78ad91a451d4237c612cf22eafc6c WatchSource:0}: Error finding container 67749dc1b4bb3dfdc39c890d1e54352654f78ad91a451d4237c612cf22eafc6c: Status 404 returned error can't find the container with id 67749dc1b4bb3dfdc39c890d1e54352654f78ad91a451d4237c612cf22eafc6c Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.506339 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5c7b91a3-10d6-465d-aff4-4855d697b54f","Type":"ContainerStarted","Data":"ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba"} Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.507455 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.508543 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tk256-config-8wrkv" event={"ID":"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68","Type":"ContainerStarted","Data":"400b8b10f6b7cc5e6032136e65cbc7e66eaf5ed5ded8ec5da99f57d7efae3bb4"} Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.508604 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tk256-config-8wrkv" event={"ID":"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68","Type":"ContainerStarted","Data":"67749dc1b4bb3dfdc39c890d1e54352654f78ad91a451d4237c612cf22eafc6c"} Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.510314 4735 generic.go:334] "Generic (PLEG): container finished" podID="c8874208-49c3-4e6b-bdb3-ef40de3d9ec7" containerID="4b531105bad682db1ce00829f8cf5b929722989c2d49e1610e1aa8140affcbce" exitCode=0 Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.510404 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-frck7" event={"ID":"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7","Type":"ContainerDied","Data":"4b531105bad682db1ce00829f8cf5b929722989c2d49e1610e1aa8140affcbce"} Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.510519 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-frck7" event={"ID":"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7","Type":"ContainerStarted","Data":"cc959deef23ce3450971377c27606f77bcd82533e21f1606035f874f18fdb408"} Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.512496 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc3c285a-910a-4cb1-968e-636216557d09","Type":"ContainerStarted","Data":"ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1"} Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.512780 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.513976 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-djbjm" event={"ID":"be509f30-9a8a-4cd2-8bf7-b8132bd92a97","Type":"ContainerStarted","Data":"6f82ddd6d3abf184aed0373376012195adbe6090c34df6d80c183d78e380a37f"} Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.533396 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.67700535 podStartE2EDuration="1m19.533380737s" podCreationTimestamp="2026-02-15 20:32:49 +0000 UTC" firstStartedPulling="2026-02-15 20:32:51.922359121 +0000 UTC m=+979.788374744" lastFinishedPulling="2026-02-15 20:33:30.778734498 +0000 UTC m=+1018.644750131" observedRunningTime="2026-02-15 20:34:08.529495882 +0000 UTC m=+1056.395511505" watchObservedRunningTime="2026-02-15 20:34:08.533380737 +0000 UTC m=+1056.399396350" Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.580446 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-djbjm" podStartSLOduration=2.542185581 podStartE2EDuration="14.580409566s" podCreationTimestamp="2026-02-15 20:33:54 +0000 UTC" firstStartedPulling="2026-02-15 20:33:55.597611598 +0000 UTC m=+1043.463627221" lastFinishedPulling="2026-02-15 20:34:07.635835583 +0000 UTC m=+1055.501851206" observedRunningTime="2026-02-15 20:34:08.578165696 +0000 UTC m=+1056.444181329" watchObservedRunningTime="2026-02-15 20:34:08.580409566 +0000 UTC m=+1056.446425179" Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.608413 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.444174405 podStartE2EDuration="1m19.608395191s" podCreationTimestamp="2026-02-15 20:32:49 +0000 UTC" firstStartedPulling="2026-02-15 20:32:51.618424437 +0000 UTC m=+979.484440060" lastFinishedPulling="2026-02-15 20:33:30.782645223 +0000 UTC m=+1018.648660846" observedRunningTime="2026-02-15 20:34:08.602452821 +0000 UTC m=+1056.468468434" watchObservedRunningTime="2026-02-15 20:34:08.608395191 +0000 UTC m=+1056.474410814" Feb 15 20:34:08 crc kubenswrapper[4735]: I0215 20:34:08.627089 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-tk256-config-8wrkv" podStartSLOduration=4.627070005 podStartE2EDuration="4.627070005s" podCreationTimestamp="2026-02-15 20:34:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:08.620361414 +0000 UTC m=+1056.486377037" watchObservedRunningTime="2026-02-15 20:34:08.627070005 +0000 UTC m=+1056.493085618" Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.522456 4735 generic.go:334] "Generic (PLEG): container finished" podID="f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" containerID="400b8b10f6b7cc5e6032136e65cbc7e66eaf5ed5ded8ec5da99f57d7efae3bb4" exitCode=0 Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.522758 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tk256-config-8wrkv" event={"ID":"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68","Type":"ContainerDied","Data":"400b8b10f6b7cc5e6032136e65cbc7e66eaf5ed5ded8ec5da99f57d7efae3bb4"} Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.533384 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-tk256" Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.856180 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-frck7" Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.856420 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.863494 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a9642a7f-e8aa-4368-b2b9-3da6537c4d8f-etc-swift\") pod \"swift-storage-0\" (UID: \"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f\") " pod="openstack/swift-storage-0" Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.958272 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc6dn\" (UniqueName: \"kubernetes.io/projected/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-kube-api-access-rc6dn\") pod \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\" (UID: \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\") " Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.958354 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-operator-scripts\") pod \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\" (UID: \"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7\") " Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.958753 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c8874208-49c3-4e6b-bdb3-ef40de3d9ec7" (UID: "c8874208-49c3-4e6b-bdb3-ef40de3d9ec7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:09 crc kubenswrapper[4735]: I0215 20:34:09.973668 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-kube-api-access-rc6dn" (OuterVolumeSpecName: "kube-api-access-rc6dn") pod "c8874208-49c3-4e6b-bdb3-ef40de3d9ec7" (UID: "c8874208-49c3-4e6b-bdb3-ef40de3d9ec7"). InnerVolumeSpecName "kube-api-access-rc6dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:10 crc kubenswrapper[4735]: I0215 20:34:10.059791 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc6dn\" (UniqueName: \"kubernetes.io/projected/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-kube-api-access-rc6dn\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:10 crc kubenswrapper[4735]: I0215 20:34:10.059827 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:10 crc kubenswrapper[4735]: I0215 20:34:10.143558 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 15 20:34:10 crc kubenswrapper[4735]: I0215 20:34:10.536531 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-frck7" event={"ID":"c8874208-49c3-4e6b-bdb3-ef40de3d9ec7","Type":"ContainerDied","Data":"cc959deef23ce3450971377c27606f77bcd82533e21f1606035f874f18fdb408"} Feb 15 20:34:10 crc kubenswrapper[4735]: I0215 20:34:10.536803 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc959deef23ce3450971377c27606f77bcd82533e21f1606035f874f18fdb408" Feb 15 20:34:10 crc kubenswrapper[4735]: I0215 20:34:10.536570 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-frck7" Feb 15 20:34:10 crc kubenswrapper[4735]: I0215 20:34:10.776843 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 15 20:34:10 crc kubenswrapper[4735]: I0215 20:34:10.903626 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.078402 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-log-ovn\") pod \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.078473 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run-ovn\") pod \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.078514 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5lc2\" (UniqueName: \"kubernetes.io/projected/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-kube-api-access-p5lc2\") pod \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.078547 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-scripts\") pod \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.078597 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run\") pod \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.078663 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-additional-scripts\") pod \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\" (UID: \"f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68\") " Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.079526 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" (UID: "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.079561 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" (UID: "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.079579 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" (UID: "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.081019 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-scripts" (OuterVolumeSpecName: "scripts") pod "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" (UID: "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.081050 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run" (OuterVolumeSpecName: "var-run") pod "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" (UID: "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.104127 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-kube-api-access-p5lc2" (OuterVolumeSpecName: "kube-api-access-p5lc2") pod "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" (UID: "f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68"). InnerVolumeSpecName "kube-api-access-p5lc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.180908 4735 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.180937 4735 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.180959 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5lc2\" (UniqueName: \"kubernetes.io/projected/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-kube-api-access-p5lc2\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.180970 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.180979 4735 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-var-run\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.180987 4735 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.229274 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tk256-config-8wrkv"] Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.237200 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tk256-config-8wrkv"] Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.339388 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tk256-config-2rvf5"] Feb 15 20:34:11 crc kubenswrapper[4735]: E0215 20:34:11.339740 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8874208-49c3-4e6b-bdb3-ef40de3d9ec7" containerName="mariadb-account-create-update" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.339752 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8874208-49c3-4e6b-bdb3-ef40de3d9ec7" containerName="mariadb-account-create-update" Feb 15 20:34:11 crc kubenswrapper[4735]: E0215 20:34:11.339771 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" containerName="ovn-config" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.339789 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" containerName="ovn-config" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.339969 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" containerName="ovn-config" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.339984 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8874208-49c3-4e6b-bdb3-ef40de3d9ec7" containerName="mariadb-account-create-update" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.340491 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.351356 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tk256-config-2rvf5"] Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.485451 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-scripts\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.485510 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.485534 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prcbn\" (UniqueName: \"kubernetes.io/projected/37e07409-4e1a-4321-b1b1-9c2b84860510-kube-api-access-prcbn\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.485771 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run-ovn\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.485837 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-log-ovn\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.485900 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-additional-scripts\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.545120 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67749dc1b4bb3dfdc39c890d1e54352654f78ad91a451d4237c612cf22eafc6c" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.545185 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256-config-8wrkv" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.550312 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"873850d8b56837f6113236a65ac7be14524e38ed83500624efd8479818de38f8"} Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.587586 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-scripts\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.587626 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.587648 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prcbn\" (UniqueName: \"kubernetes.io/projected/37e07409-4e1a-4321-b1b1-9c2b84860510-kube-api-access-prcbn\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.587704 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run-ovn\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.587730 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-log-ovn\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.587758 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-additional-scripts\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.588073 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.588104 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-log-ovn\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.588178 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run-ovn\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.588411 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-additional-scripts\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.592542 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-scripts\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.618915 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prcbn\" (UniqueName: \"kubernetes.io/projected/37e07409-4e1a-4321-b1b1-9c2b84860510-kube-api-access-prcbn\") pod \"ovn-controller-tk256-config-2rvf5\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:11 crc kubenswrapper[4735]: I0215 20:34:11.656896 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:12 crc kubenswrapper[4735]: I0215 20:34:12.349106 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tk256-config-2rvf5"] Feb 15 20:34:12 crc kubenswrapper[4735]: W0215 20:34:12.353927 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37e07409_4e1a_4321_b1b1_9c2b84860510.slice/crio-83d63a0cbcc99be0d871ab28c4176966dabaca4147dcb4e96d39772b55822e3f WatchSource:0}: Error finding container 83d63a0cbcc99be0d871ab28c4176966dabaca4147dcb4e96d39772b55822e3f: Status 404 returned error can't find the container with id 83d63a0cbcc99be0d871ab28c4176966dabaca4147dcb4e96d39772b55822e3f Feb 15 20:34:12 crc kubenswrapper[4735]: I0215 20:34:12.570975 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tk256-config-2rvf5" event={"ID":"37e07409-4e1a-4321-b1b1-9c2b84860510","Type":"ContainerStarted","Data":"83d63a0cbcc99be0d871ab28c4176966dabaca4147dcb4e96d39772b55822e3f"} Feb 15 20:34:12 crc kubenswrapper[4735]: I0215 20:34:12.574040 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"3f3170a2cecfface296aaeb1dfa7d804da50bb180f233ffa42c9f761f9954e08"} Feb 15 20:34:12 crc kubenswrapper[4735]: I0215 20:34:12.574081 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"e3d4203699b417ad5beb732afab7776b5290ac12bf9be2a8bbc60db377cca7bd"} Feb 15 20:34:12 crc kubenswrapper[4735]: I0215 20:34:12.574093 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"0e7e64caa81fe9e1dc054bf60e86e3075bc4312aa1cb6c5942852eb068134895"} Feb 15 20:34:12 crc kubenswrapper[4735]: I0215 20:34:12.896148 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68" path="/var/lib/kubelet/pods/f1bbe9fa-32f5-4a8e-8ba8-39f88fffbc68/volumes" Feb 15 20:34:13 crc kubenswrapper[4735]: I0215 20:34:13.581662 4735 generic.go:334] "Generic (PLEG): container finished" podID="37e07409-4e1a-4321-b1b1-9c2b84860510" containerID="1bfc08d2407ef98f0c67d479ee227ecbe94d6bfa9ac6f4963f0cdc56de584c08" exitCode=0 Feb 15 20:34:13 crc kubenswrapper[4735]: I0215 20:34:13.581822 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tk256-config-2rvf5" event={"ID":"37e07409-4e1a-4321-b1b1-9c2b84860510","Type":"ContainerDied","Data":"1bfc08d2407ef98f0c67d479ee227ecbe94d6bfa9ac6f4963f0cdc56de584c08"} Feb 15 20:34:13 crc kubenswrapper[4735]: I0215 20:34:13.586364 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"38d72bfb8c5a56ead9f27b775d556a81d6a5f681261e1dec16d374593229ac75"} Feb 15 20:34:14 crc kubenswrapper[4735]: I0215 20:34:14.597061 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"83d84960bbbc9c1de0078681cac96604017acce0c853743db8a5d890971c9f84"} Feb 15 20:34:14 crc kubenswrapper[4735]: I0215 20:34:14.597466 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"8f30da0a8ea018db9a717cb85a598a911b90058e2d597a945a29e07157705ec2"} Feb 15 20:34:14 crc kubenswrapper[4735]: I0215 20:34:14.597483 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"be07c72d2b5c0e457cee45381645c52d2d8237ece572717862de09561fc3bc5d"} Feb 15 20:34:14 crc kubenswrapper[4735]: I0215 20:34:14.860895 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.044630 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run-ovn\") pod \"37e07409-4e1a-4321-b1b1-9c2b84860510\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.044780 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-additional-scripts\") pod \"37e07409-4e1a-4321-b1b1-9c2b84860510\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.044836 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-scripts\") pod \"37e07409-4e1a-4321-b1b1-9c2b84860510\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.044867 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-log-ovn\") pod \"37e07409-4e1a-4321-b1b1-9c2b84860510\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.044901 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run\") pod \"37e07409-4e1a-4321-b1b1-9c2b84860510\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.044930 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prcbn\" (UniqueName: \"kubernetes.io/projected/37e07409-4e1a-4321-b1b1-9c2b84860510-kube-api-access-prcbn\") pod \"37e07409-4e1a-4321-b1b1-9c2b84860510\" (UID: \"37e07409-4e1a-4321-b1b1-9c2b84860510\") " Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.045178 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "37e07409-4e1a-4321-b1b1-9c2b84860510" (UID: "37e07409-4e1a-4321-b1b1-9c2b84860510"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.045240 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run" (OuterVolumeSpecName: "var-run") pod "37e07409-4e1a-4321-b1b1-9c2b84860510" (UID: "37e07409-4e1a-4321-b1b1-9c2b84860510"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.045269 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "37e07409-4e1a-4321-b1b1-9c2b84860510" (UID: "37e07409-4e1a-4321-b1b1-9c2b84860510"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.045844 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "37e07409-4e1a-4321-b1b1-9c2b84860510" (UID: "37e07409-4e1a-4321-b1b1-9c2b84860510"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.046188 4735 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.046209 4735 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.046217 4735 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/37e07409-4e1a-4321-b1b1-9c2b84860510-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.046225 4735 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.046371 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-scripts" (OuterVolumeSpecName: "scripts") pod "37e07409-4e1a-4321-b1b1-9c2b84860510" (UID: "37e07409-4e1a-4321-b1b1-9c2b84860510"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.062135 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37e07409-4e1a-4321-b1b1-9c2b84860510-kube-api-access-prcbn" (OuterVolumeSpecName: "kube-api-access-prcbn") pod "37e07409-4e1a-4321-b1b1-9c2b84860510" (UID: "37e07409-4e1a-4321-b1b1-9c2b84860510"). InnerVolumeSpecName "kube-api-access-prcbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.149057 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37e07409-4e1a-4321-b1b1-9c2b84860510-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.149086 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prcbn\" (UniqueName: \"kubernetes.io/projected/37e07409-4e1a-4321-b1b1-9c2b84860510-kube-api-access-prcbn\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.607408 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tk256-config-2rvf5" event={"ID":"37e07409-4e1a-4321-b1b1-9c2b84860510","Type":"ContainerDied","Data":"83d63a0cbcc99be0d871ab28c4176966dabaca4147dcb4e96d39772b55822e3f"} Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.608728 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83d63a0cbcc99be0d871ab28c4176966dabaca4147dcb4e96d39772b55822e3f" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.607721 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tk256-config-2rvf5" Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.610888 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"b29809d648e160eb7367cdc6c15f92e97e2607ef7494cf912220b1255fa76710"} Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.946223 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tk256-config-2rvf5"] Feb 15 20:34:15 crc kubenswrapper[4735]: I0215 20:34:15.971956 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tk256-config-2rvf5"] Feb 15 20:34:16 crc kubenswrapper[4735]: I0215 20:34:16.747277 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"fb780c3af7721813fceb711493c9411f0643f7fdd5d0338c5752b9ca3687a5eb"} Feb 15 20:34:16 crc kubenswrapper[4735]: I0215 20:34:16.747328 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"70dcfcedd81a43e22424d9ddbc3775f64339ff18b6d2dcb9434d522a6fb5bce8"} Feb 15 20:34:16 crc kubenswrapper[4735]: I0215 20:34:16.899742 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37e07409-4e1a-4321-b1b1-9c2b84860510" path="/var/lib/kubelet/pods/37e07409-4e1a-4321-b1b1-9c2b84860510/volumes" Feb 15 20:34:17 crc kubenswrapper[4735]: I0215 20:34:17.757272 4735 generic.go:334] "Generic (PLEG): container finished" podID="be509f30-9a8a-4cd2-8bf7-b8132bd92a97" containerID="6f82ddd6d3abf184aed0373376012195adbe6090c34df6d80c183d78e380a37f" exitCode=0 Feb 15 20:34:17 crc kubenswrapper[4735]: I0215 20:34:17.757557 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-djbjm" event={"ID":"be509f30-9a8a-4cd2-8bf7-b8132bd92a97","Type":"ContainerDied","Data":"6f82ddd6d3abf184aed0373376012195adbe6090c34df6d80c183d78e380a37f"} Feb 15 20:34:17 crc kubenswrapper[4735]: I0215 20:34:17.763272 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"dac04f68d2f4de6393f6d780713342296360e53a60aefd8fbec3f45860d2b27f"} Feb 15 20:34:17 crc kubenswrapper[4735]: I0215 20:34:17.763300 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"c381e7f6c6ebfb25f61400ad0f54588ffaec82dbc5a046ad9902de7fb7438b4d"} Feb 15 20:34:17 crc kubenswrapper[4735]: I0215 20:34:17.763312 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"4b4ec52ad2514470eb93e5b0a71df742abd6404533ee996a915a9b3d96fdd887"} Feb 15 20:34:17 crc kubenswrapper[4735]: I0215 20:34:17.763320 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"da01464d325bb65ed8c5eef560e9f900045a926ee30e64687dce0427a13a8d68"} Feb 15 20:34:17 crc kubenswrapper[4735]: I0215 20:34:17.763329 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a9642a7f-e8aa-4368-b2b9-3da6537c4d8f","Type":"ContainerStarted","Data":"066877cb6bdc5afa5004eaa59699e907d8e98e65a849d02a77610a9f12a640ad"} Feb 15 20:34:17 crc kubenswrapper[4735]: I0215 20:34:17.822485 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.029514133 podStartE2EDuration="41.822469837s" podCreationTimestamp="2026-02-15 20:33:36 +0000 UTC" firstStartedPulling="2026-02-15 20:34:10.778239648 +0000 UTC m=+1058.644255281" lastFinishedPulling="2026-02-15 20:34:15.571195362 +0000 UTC m=+1063.437210985" observedRunningTime="2026-02-15 20:34:17.817848242 +0000 UTC m=+1065.683863865" watchObservedRunningTime="2026-02-15 20:34:17.822469837 +0000 UTC m=+1065.688485460" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.116755 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-nzz9h"] Feb 15 20:34:18 crc kubenswrapper[4735]: E0215 20:34:18.117164 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e07409-4e1a-4321-b1b1-9c2b84860510" containerName="ovn-config" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.117185 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e07409-4e1a-4321-b1b1-9c2b84860510" containerName="ovn-config" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.117388 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="37e07409-4e1a-4321-b1b1-9c2b84860510" containerName="ovn-config" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.118378 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.120132 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.131001 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-nzz9h"] Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.263549 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.263612 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.263691 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.263764 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.263805 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-config\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.263833 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whxd8\" (UniqueName: \"kubernetes.io/projected/236bc253-c895-44c1-879f-48362a006c48-kube-api-access-whxd8\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.365095 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.365188 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.365254 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.365307 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.365340 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-config\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.365365 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whxd8\" (UniqueName: \"kubernetes.io/projected/236bc253-c895-44c1-879f-48362a006c48-kube-api-access-whxd8\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.366353 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.366426 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.366546 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.366655 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-config\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.367305 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.408807 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whxd8\" (UniqueName: \"kubernetes.io/projected/236bc253-c895-44c1-879f-48362a006c48-kube-api-access-whxd8\") pod \"dnsmasq-dns-5c79d794d7-nzz9h\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.434849 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:18 crc kubenswrapper[4735]: I0215 20:34:18.969245 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-nzz9h"] Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.155375 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-djbjm" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.286762 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnqnl\" (UniqueName: \"kubernetes.io/projected/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-kube-api-access-fnqnl\") pod \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.286817 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-db-sync-config-data\") pod \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.286912 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-config-data\") pod \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.287047 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-combined-ca-bundle\") pod \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\" (UID: \"be509f30-9a8a-4cd2-8bf7-b8132bd92a97\") " Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.293557 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-kube-api-access-fnqnl" (OuterVolumeSpecName: "kube-api-access-fnqnl") pod "be509f30-9a8a-4cd2-8bf7-b8132bd92a97" (UID: "be509f30-9a8a-4cd2-8bf7-b8132bd92a97"). InnerVolumeSpecName "kube-api-access-fnqnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.306917 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "be509f30-9a8a-4cd2-8bf7-b8132bd92a97" (UID: "be509f30-9a8a-4cd2-8bf7-b8132bd92a97"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.315112 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be509f30-9a8a-4cd2-8bf7-b8132bd92a97" (UID: "be509f30-9a8a-4cd2-8bf7-b8132bd92a97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.329093 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-config-data" (OuterVolumeSpecName: "config-data") pod "be509f30-9a8a-4cd2-8bf7-b8132bd92a97" (UID: "be509f30-9a8a-4cd2-8bf7-b8132bd92a97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.388739 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.388774 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnqnl\" (UniqueName: \"kubernetes.io/projected/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-kube-api-access-fnqnl\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.388785 4735 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.388793 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be509f30-9a8a-4cd2-8bf7-b8132bd92a97-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.783150 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" event={"ID":"236bc253-c895-44c1-879f-48362a006c48","Type":"ContainerStarted","Data":"c0f92c36a30cbf3df67937df1ce2e12832d40a8a3ae60d81bf0fc958fc7896b7"} Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.787296 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-djbjm" event={"ID":"be509f30-9a8a-4cd2-8bf7-b8132bd92a97","Type":"ContainerDied","Data":"2a9d1850d4c5ac17736735464e117a308cd9d9c276819788c57eaed404806951"} Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.787343 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a9d1850d4c5ac17736735464e117a308cd9d9c276819788c57eaed404806951" Feb 15 20:34:19 crc kubenswrapper[4735]: I0215 20:34:19.787434 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-djbjm" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.193114 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-nzz9h"] Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.237851 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-dfdmw"] Feb 15 20:34:20 crc kubenswrapper[4735]: E0215 20:34:20.238216 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be509f30-9a8a-4cd2-8bf7-b8132bd92a97" containerName="glance-db-sync" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.238234 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="be509f30-9a8a-4cd2-8bf7-b8132bd92a97" containerName="glance-db-sync" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.238401 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="be509f30-9a8a-4cd2-8bf7-b8132bd92a97" containerName="glance-db-sync" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.239177 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.258254 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-dfdmw"] Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.423308 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.423623 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.423646 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b852\" (UniqueName: \"kubernetes.io/projected/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-kube-api-access-6b852\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.423671 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-config\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.423692 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.423722 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.525570 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.526653 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.527315 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.527349 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.527370 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b852\" (UniqueName: \"kubernetes.io/projected/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-kube-api-access-6b852\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.527394 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-config\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.527211 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.526588 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.528294 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.528509 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.528575 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-config\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.545159 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b852\" (UniqueName: \"kubernetes.io/projected/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-kube-api-access-6b852\") pod \"dnsmasq-dns-5f59b8f679-dfdmw\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.553695 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.797278 4735 generic.go:334] "Generic (PLEG): container finished" podID="236bc253-c895-44c1-879f-48362a006c48" containerID="eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438" exitCode=0 Feb 15 20:34:20 crc kubenswrapper[4735]: I0215 20:34:20.797404 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" event={"ID":"236bc253-c895-44c1-879f-48362a006c48","Type":"ContainerDied","Data":"eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438"} Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.035511 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-dfdmw"] Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.080233 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="dc3c285a-910a-4cb1-968e-636216557d09" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.363198 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.805209 4735 generic.go:334] "Generic (PLEG): container finished" podID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerID="b6ac5646db598059deeb27c0b0edbf68ee2c5051dc54d264c2fdff563aceecb9" exitCode=0 Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.805268 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" event={"ID":"fc2f480c-f0b3-4677-ad26-2b4b5071ad44","Type":"ContainerDied","Data":"b6ac5646db598059deeb27c0b0edbf68ee2c5051dc54d264c2fdff563aceecb9"} Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.805293 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" event={"ID":"fc2f480c-f0b3-4677-ad26-2b4b5071ad44","Type":"ContainerStarted","Data":"4838722be565f495ceac73db949cb7d656d47af2da5a40254006801cf26db1d6"} Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.807407 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" event={"ID":"236bc253-c895-44c1-879f-48362a006c48","Type":"ContainerStarted","Data":"a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf"} Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.807494 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" podUID="236bc253-c895-44c1-879f-48362a006c48" containerName="dnsmasq-dns" containerID="cri-o://a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf" gracePeriod=10 Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.807534 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:21 crc kubenswrapper[4735]: I0215 20:34:21.875858 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" podStartSLOduration=3.875841771 podStartE2EDuration="3.875841771s" podCreationTimestamp="2026-02-15 20:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:21.871098832 +0000 UTC m=+1069.737114465" watchObservedRunningTime="2026-02-15 20:34:21.875841771 +0000 UTC m=+1069.741857394" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.223130 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.359080 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-swift-storage-0\") pod \"236bc253-c895-44c1-879f-48362a006c48\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.359144 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-svc\") pod \"236bc253-c895-44c1-879f-48362a006c48\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.359215 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-config\") pod \"236bc253-c895-44c1-879f-48362a006c48\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.359253 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whxd8\" (UniqueName: \"kubernetes.io/projected/236bc253-c895-44c1-879f-48362a006c48-kube-api-access-whxd8\") pod \"236bc253-c895-44c1-879f-48362a006c48\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.359287 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-sb\") pod \"236bc253-c895-44c1-879f-48362a006c48\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.359368 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-nb\") pod \"236bc253-c895-44c1-879f-48362a006c48\" (UID: \"236bc253-c895-44c1-879f-48362a006c48\") " Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.368166 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/236bc253-c895-44c1-879f-48362a006c48-kube-api-access-whxd8" (OuterVolumeSpecName: "kube-api-access-whxd8") pod "236bc253-c895-44c1-879f-48362a006c48" (UID: "236bc253-c895-44c1-879f-48362a006c48"). InnerVolumeSpecName "kube-api-access-whxd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.405057 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "236bc253-c895-44c1-879f-48362a006c48" (UID: "236bc253-c895-44c1-879f-48362a006c48"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.406232 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-config" (OuterVolumeSpecName: "config") pod "236bc253-c895-44c1-879f-48362a006c48" (UID: "236bc253-c895-44c1-879f-48362a006c48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.410703 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "236bc253-c895-44c1-879f-48362a006c48" (UID: "236bc253-c895-44c1-879f-48362a006c48"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.423569 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "236bc253-c895-44c1-879f-48362a006c48" (UID: "236bc253-c895-44c1-879f-48362a006c48"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.432833 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "236bc253-c895-44c1-879f-48362a006c48" (UID: "236bc253-c895-44c1-879f-48362a006c48"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.460856 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.460896 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.460906 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.460917 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.460925 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whxd8\" (UniqueName: \"kubernetes.io/projected/236bc253-c895-44c1-879f-48362a006c48-kube-api-access-whxd8\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.460935 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/236bc253-c895-44c1-879f-48362a006c48-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.815381 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" event={"ID":"fc2f480c-f0b3-4677-ad26-2b4b5071ad44","Type":"ContainerStarted","Data":"6d9cdedf8bcecd556655db096a7f647888ea55cf1a0541f7f0672b100320fec3"} Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.815691 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.817519 4735 generic.go:334] "Generic (PLEG): container finished" podID="236bc253-c895-44c1-879f-48362a006c48" containerID="a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf" exitCode=0 Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.817547 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" event={"ID":"236bc253-c895-44c1-879f-48362a006c48","Type":"ContainerDied","Data":"a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf"} Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.817552 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.817564 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-nzz9h" event={"ID":"236bc253-c895-44c1-879f-48362a006c48","Type":"ContainerDied","Data":"c0f92c36a30cbf3df67937df1ce2e12832d40a8a3ae60d81bf0fc958fc7896b7"} Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.817581 4735 scope.go:117] "RemoveContainer" containerID="a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.837610 4735 scope.go:117] "RemoveContainer" containerID="eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.863853 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" podStartSLOduration=2.863827334 podStartE2EDuration="2.863827334s" podCreationTimestamp="2026-02-15 20:34:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:22.849336873 +0000 UTC m=+1070.715352496" watchObservedRunningTime="2026-02-15 20:34:22.863827334 +0000 UTC m=+1070.729842967" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.867656 4735 scope.go:117] "RemoveContainer" containerID="a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf" Feb 15 20:34:22 crc kubenswrapper[4735]: E0215 20:34:22.868270 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf\": container with ID starting with a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf not found: ID does not exist" containerID="a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.868302 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf"} err="failed to get container status \"a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf\": rpc error: code = NotFound desc = could not find container \"a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf\": container with ID starting with a6f87ba84f75dd052b5d94fad7ae744c71ad03de0197f103a542a52e505506cf not found: ID does not exist" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.868320 4735 scope.go:117] "RemoveContainer" containerID="eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438" Feb 15 20:34:22 crc kubenswrapper[4735]: E0215 20:34:22.868643 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438\": container with ID starting with eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438 not found: ID does not exist" containerID="eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.868667 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438"} err="failed to get container status \"eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438\": rpc error: code = NotFound desc = could not find container \"eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438\": container with ID starting with eed53d51077475cf6cf364cb114bf413fe20dedeeb2002615c559076e4747438 not found: ID does not exist" Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.881893 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-nzz9h"] Feb 15 20:34:22 crc kubenswrapper[4735]: I0215 20:34:22.905696 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-nzz9h"] Feb 15 20:34:24 crc kubenswrapper[4735]: I0215 20:34:24.902058 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="236bc253-c895-44c1-879f-48362a006c48" path="/var/lib/kubelet/pods/236bc253-c895-44c1-879f-48362a006c48/volumes" Feb 15 20:34:30 crc kubenswrapper[4735]: I0215 20:34:30.555829 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:34:30 crc kubenswrapper[4735]: I0215 20:34:30.635506 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-k7qln"] Feb 15 20:34:30 crc kubenswrapper[4735]: I0215 20:34:30.635838 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" podUID="d482296c-b6fe-46a1-8798-696c4d06cdf0" containerName="dnsmasq-dns" containerID="cri-o://d8280a3ef47c7e342bea709f44e836795ab36a022326f5c00708516014e83e20" gracePeriod=10 Feb 15 20:34:30 crc kubenswrapper[4735]: I0215 20:34:30.903431 4735 generic.go:334] "Generic (PLEG): container finished" podID="d482296c-b6fe-46a1-8798-696c4d06cdf0" containerID="d8280a3ef47c7e342bea709f44e836795ab36a022326f5c00708516014e83e20" exitCode=0 Feb 15 20:34:30 crc kubenswrapper[4735]: I0215 20:34:30.903705 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" event={"ID":"d482296c-b6fe-46a1-8798-696c4d06cdf0","Type":"ContainerDied","Data":"d8280a3ef47c7e342bea709f44e836795ab36a022326f5c00708516014e83e20"} Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.073587 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.132832 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.217124 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-nb\") pod \"d482296c-b6fe-46a1-8798-696c4d06cdf0\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.217211 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-sb\") pod \"d482296c-b6fe-46a1-8798-696c4d06cdf0\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.217264 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkd58\" (UniqueName: \"kubernetes.io/projected/d482296c-b6fe-46a1-8798-696c4d06cdf0-kube-api-access-xkd58\") pod \"d482296c-b6fe-46a1-8798-696c4d06cdf0\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.217324 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-config\") pod \"d482296c-b6fe-46a1-8798-696c4d06cdf0\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.217366 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-dns-svc\") pod \"d482296c-b6fe-46a1-8798-696c4d06cdf0\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.230152 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d482296c-b6fe-46a1-8798-696c4d06cdf0-kube-api-access-xkd58" (OuterVolumeSpecName: "kube-api-access-xkd58") pod "d482296c-b6fe-46a1-8798-696c4d06cdf0" (UID: "d482296c-b6fe-46a1-8798-696c4d06cdf0"). InnerVolumeSpecName "kube-api-access-xkd58". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.348534 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d482296c-b6fe-46a1-8798-696c4d06cdf0" (UID: "d482296c-b6fe-46a1-8798-696c4d06cdf0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.349892 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-dns-svc\") pod \"d482296c-b6fe-46a1-8798-696c4d06cdf0\" (UID: \"d482296c-b6fe-46a1-8798-696c4d06cdf0\") " Feb 15 20:34:31 crc kubenswrapper[4735]: W0215 20:34:31.350077 4735 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/d482296c-b6fe-46a1-8798-696c4d06cdf0/volumes/kubernetes.io~configmap/dns-svc Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.350094 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d482296c-b6fe-46a1-8798-696c4d06cdf0" (UID: "d482296c-b6fe-46a1-8798-696c4d06cdf0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.350686 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkd58\" (UniqueName: \"kubernetes.io/projected/d482296c-b6fe-46a1-8798-696c4d06cdf0-kube-api-access-xkd58\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.350702 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.396541 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d482296c-b6fe-46a1-8798-696c4d06cdf0" (UID: "d482296c-b6fe-46a1-8798-696c4d06cdf0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.409163 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-config" (OuterVolumeSpecName: "config") pod "d482296c-b6fe-46a1-8798-696c4d06cdf0" (UID: "d482296c-b6fe-46a1-8798-696c4d06cdf0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.421357 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d482296c-b6fe-46a1-8798-696c4d06cdf0" (UID: "d482296c-b6fe-46a1-8798-696c4d06cdf0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.451998 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.452195 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.452207 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d482296c-b6fe-46a1-8798-696c4d06cdf0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.474297 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-j6mch"] Feb 15 20:34:31 crc kubenswrapper[4735]: E0215 20:34:31.474592 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236bc253-c895-44c1-879f-48362a006c48" containerName="init" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.474607 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="236bc253-c895-44c1-879f-48362a006c48" containerName="init" Feb 15 20:34:31 crc kubenswrapper[4735]: E0215 20:34:31.474628 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d482296c-b6fe-46a1-8798-696c4d06cdf0" containerName="dnsmasq-dns" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.474634 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d482296c-b6fe-46a1-8798-696c4d06cdf0" containerName="dnsmasq-dns" Feb 15 20:34:31 crc kubenswrapper[4735]: E0215 20:34:31.474642 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d482296c-b6fe-46a1-8798-696c4d06cdf0" containerName="init" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.474648 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d482296c-b6fe-46a1-8798-696c4d06cdf0" containerName="init" Feb 15 20:34:31 crc kubenswrapper[4735]: E0215 20:34:31.474658 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236bc253-c895-44c1-879f-48362a006c48" containerName="dnsmasq-dns" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.474663 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="236bc253-c895-44c1-879f-48362a006c48" containerName="dnsmasq-dns" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.474821 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="236bc253-c895-44c1-879f-48362a006c48" containerName="dnsmasq-dns" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.474845 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d482296c-b6fe-46a1-8798-696c4d06cdf0" containerName="dnsmasq-dns" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.475341 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.493495 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-j6mch"] Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.553801 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-operator-scripts\") pod \"cinder-db-create-j6mch\" (UID: \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\") " pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.553894 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zhvb\" (UniqueName: \"kubernetes.io/projected/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-kube-api-access-5zhvb\") pod \"cinder-db-create-j6mch\" (UID: \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\") " pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.654621 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-operator-scripts\") pod \"cinder-db-create-j6mch\" (UID: \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\") " pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.655631 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zhvb\" (UniqueName: \"kubernetes.io/projected/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-kube-api-access-5zhvb\") pod \"cinder-db-create-j6mch\" (UID: \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\") " pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.655342 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-operator-scripts\") pod \"cinder-db-create-j6mch\" (UID: \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\") " pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.675562 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-fd5e-account-create-update-m58kx"] Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.676491 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.679070 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zhvb\" (UniqueName: \"kubernetes.io/projected/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-kube-api-access-5zhvb\") pod \"cinder-db-create-j6mch\" (UID: \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\") " pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.680035 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.699993 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fd5e-account-create-update-m58kx"] Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.757308 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f82da625-890e-43b7-8dee-7ab76e375b52-operator-scripts\") pod \"barbican-fd5e-account-create-update-m58kx\" (UID: \"f82da625-890e-43b7-8dee-7ab76e375b52\") " pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.757448 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flxg4\" (UniqueName: \"kubernetes.io/projected/f82da625-890e-43b7-8dee-7ab76e375b52-kube-api-access-flxg4\") pod \"barbican-fd5e-account-create-update-m58kx\" (UID: \"f82da625-890e-43b7-8dee-7ab76e375b52\") " pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.772543 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-d27lr"] Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.773435 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.781158 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d27lr"] Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.799882 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.804897 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f244-account-create-update-kc8sg"] Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.805875 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.807657 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.823015 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f244-account-create-update-kc8sg"] Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.858465 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7qbv\" (UniqueName: \"kubernetes.io/projected/d3bc62da-42b0-4f89-907a-64b5d2443fcf-kube-api-access-j7qbv\") pod \"cinder-f244-account-create-update-kc8sg\" (UID: \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\") " pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.858745 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bc62da-42b0-4f89-907a-64b5d2443fcf-operator-scripts\") pod \"cinder-f244-account-create-update-kc8sg\" (UID: \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\") " pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.858829 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flxg4\" (UniqueName: \"kubernetes.io/projected/f82da625-890e-43b7-8dee-7ab76e375b52-kube-api-access-flxg4\") pod \"barbican-fd5e-account-create-update-m58kx\" (UID: \"f82da625-890e-43b7-8dee-7ab76e375b52\") " pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.858920 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4lj2\" (UniqueName: \"kubernetes.io/projected/36b69cc5-60b4-4560-9a60-27db0295a1cc-kube-api-access-c4lj2\") pod \"barbican-db-create-d27lr\" (UID: \"36b69cc5-60b4-4560-9a60-27db0295a1cc\") " pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.859025 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f82da625-890e-43b7-8dee-7ab76e375b52-operator-scripts\") pod \"barbican-fd5e-account-create-update-m58kx\" (UID: \"f82da625-890e-43b7-8dee-7ab76e375b52\") " pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.859127 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36b69cc5-60b4-4560-9a60-27db0295a1cc-operator-scripts\") pod \"barbican-db-create-d27lr\" (UID: \"36b69cc5-60b4-4560-9a60-27db0295a1cc\") " pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.860123 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f82da625-890e-43b7-8dee-7ab76e375b52-operator-scripts\") pod \"barbican-fd5e-account-create-update-m58kx\" (UID: \"f82da625-890e-43b7-8dee-7ab76e375b52\") " pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.893711 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flxg4\" (UniqueName: \"kubernetes.io/projected/f82da625-890e-43b7-8dee-7ab76e375b52-kube-api-access-flxg4\") pod \"barbican-fd5e-account-create-update-m58kx\" (UID: \"f82da625-890e-43b7-8dee-7ab76e375b52\") " pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.921031 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" event={"ID":"d482296c-b6fe-46a1-8798-696c4d06cdf0","Type":"ContainerDied","Data":"44bf1115d9e7a2fb2408bdcefa715230b851b1df5625f771410b911c5ca95ebd"} Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.921397 4735 scope.go:117] "RemoveContainer" containerID="d8280a3ef47c7e342bea709f44e836795ab36a022326f5c00708516014e83e20" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.921645 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-k7qln" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.960491 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7qbv\" (UniqueName: \"kubernetes.io/projected/d3bc62da-42b0-4f89-907a-64b5d2443fcf-kube-api-access-j7qbv\") pod \"cinder-f244-account-create-update-kc8sg\" (UID: \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\") " pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.960602 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bc62da-42b0-4f89-907a-64b5d2443fcf-operator-scripts\") pod \"cinder-f244-account-create-update-kc8sg\" (UID: \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\") " pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.960708 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4lj2\" (UniqueName: \"kubernetes.io/projected/36b69cc5-60b4-4560-9a60-27db0295a1cc-kube-api-access-c4lj2\") pod \"barbican-db-create-d27lr\" (UID: \"36b69cc5-60b4-4560-9a60-27db0295a1cc\") " pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.960856 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36b69cc5-60b4-4560-9a60-27db0295a1cc-operator-scripts\") pod \"barbican-db-create-d27lr\" (UID: \"36b69cc5-60b4-4560-9a60-27db0295a1cc\") " pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.961491 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36b69cc5-60b4-4560-9a60-27db0295a1cc-operator-scripts\") pod \"barbican-db-create-d27lr\" (UID: \"36b69cc5-60b4-4560-9a60-27db0295a1cc\") " pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.962474 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bc62da-42b0-4f89-907a-64b5d2443fcf-operator-scripts\") pod \"cinder-f244-account-create-update-kc8sg\" (UID: \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\") " pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.962872 4735 scope.go:117] "RemoveContainer" containerID="b8fa7ff7182371f53f58a31bc8ce267947d2a87a704042809da492950fc31605" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.977723 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ft5v7"] Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.978888 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.982322 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.982613 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.982796 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dljv4" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.983074 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.990982 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ft5v7"] Feb 15 20:34:31 crc kubenswrapper[4735]: I0215 20:34:31.991462 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.006100 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4lj2\" (UniqueName: \"kubernetes.io/projected/36b69cc5-60b4-4560-9a60-27db0295a1cc-kube-api-access-c4lj2\") pod \"barbican-db-create-d27lr\" (UID: \"36b69cc5-60b4-4560-9a60-27db0295a1cc\") " pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.010755 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-k7qln"] Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.014223 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7qbv\" (UniqueName: \"kubernetes.io/projected/d3bc62da-42b0-4f89-907a-64b5d2443fcf-kube-api-access-j7qbv\") pod \"cinder-f244-account-create-update-kc8sg\" (UID: \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\") " pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.021923 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-k7qln"] Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.114415 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-p2mmq"] Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.115353 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.119602 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-p2mmq"] Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.163854 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhrcf\" (UniqueName: \"kubernetes.io/projected/5c3905d6-692c-4791-84b0-58d7f6372642-kube-api-access-fhrcf\") pod \"keystone-db-sync-ft5v7\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.163990 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-config-data\") pod \"keystone-db-sync-ft5v7\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.164016 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-combined-ca-bundle\") pod \"keystone-db-sync-ft5v7\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.194029 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8ea4-account-create-update-sjn94"] Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.194875 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.198654 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.231418 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8ea4-account-create-update-sjn94"] Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.259386 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.268772 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-config-data\") pod \"keystone-db-sync-ft5v7\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.268812 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-combined-ca-bundle\") pod \"keystone-db-sync-ft5v7\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.268920 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqphw\" (UniqueName: \"kubernetes.io/projected/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-kube-api-access-vqphw\") pod \"neutron-db-create-p2mmq\" (UID: \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\") " pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.268969 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhrcf\" (UniqueName: \"kubernetes.io/projected/5c3905d6-692c-4791-84b0-58d7f6372642-kube-api-access-fhrcf\") pod \"keystone-db-sync-ft5v7\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.268990 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-operator-scripts\") pod \"neutron-db-create-p2mmq\" (UID: \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\") " pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.273704 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.277559 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-config-data\") pod \"keystone-db-sync-ft5v7\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.283093 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-combined-ca-bundle\") pod \"keystone-db-sync-ft5v7\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.289689 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-j6mch"] Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.298475 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhrcf\" (UniqueName: \"kubernetes.io/projected/5c3905d6-692c-4791-84b0-58d7f6372642-kube-api-access-fhrcf\") pod \"keystone-db-sync-ft5v7\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: W0215 20:34:32.312835 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2aacebd_28b2_43f7_afaa_966bd5bf7ade.slice/crio-7eb3fa960370acd110e97c605d010c4d83d9bcf85da3447335036c96f72665dc WatchSource:0}: Error finding container 7eb3fa960370acd110e97c605d010c4d83d9bcf85da3447335036c96f72665dc: Status 404 returned error can't find the container with id 7eb3fa960370acd110e97c605d010c4d83d9bcf85da3447335036c96f72665dc Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.372812 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqphw\" (UniqueName: \"kubernetes.io/projected/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-kube-api-access-vqphw\") pod \"neutron-db-create-p2mmq\" (UID: \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\") " pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.372861 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-operator-scripts\") pod \"neutron-db-create-p2mmq\" (UID: \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\") " pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.372913 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k22pr\" (UniqueName: \"kubernetes.io/projected/b66e376f-3330-4024-a409-c6820011d354-kube-api-access-k22pr\") pod \"neutron-8ea4-account-create-update-sjn94\" (UID: \"b66e376f-3330-4024-a409-c6820011d354\") " pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.372991 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66e376f-3330-4024-a409-c6820011d354-operator-scripts\") pod \"neutron-8ea4-account-create-update-sjn94\" (UID: \"b66e376f-3330-4024-a409-c6820011d354\") " pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.373875 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-operator-scripts\") pod \"neutron-db-create-p2mmq\" (UID: \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\") " pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.379829 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.388171 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fd5e-account-create-update-m58kx"] Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.395389 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqphw\" (UniqueName: \"kubernetes.io/projected/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-kube-api-access-vqphw\") pod \"neutron-db-create-p2mmq\" (UID: \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\") " pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.432346 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.474516 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k22pr\" (UniqueName: \"kubernetes.io/projected/b66e376f-3330-4024-a409-c6820011d354-kube-api-access-k22pr\") pod \"neutron-8ea4-account-create-update-sjn94\" (UID: \"b66e376f-3330-4024-a409-c6820011d354\") " pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.474602 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66e376f-3330-4024-a409-c6820011d354-operator-scripts\") pod \"neutron-8ea4-account-create-update-sjn94\" (UID: \"b66e376f-3330-4024-a409-c6820011d354\") " pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.475192 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66e376f-3330-4024-a409-c6820011d354-operator-scripts\") pod \"neutron-8ea4-account-create-update-sjn94\" (UID: \"b66e376f-3330-4024-a409-c6820011d354\") " pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.536623 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k22pr\" (UniqueName: \"kubernetes.io/projected/b66e376f-3330-4024-a409-c6820011d354-kube-api-access-k22pr\") pod \"neutron-8ea4-account-create-update-sjn94\" (UID: \"b66e376f-3330-4024-a409-c6820011d354\") " pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:32 crc kubenswrapper[4735]: I0215 20:34:32.813289 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:32 crc kubenswrapper[4735]: W0215 20:34:32.970923 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3bc62da_42b0_4f89_907a_64b5d2443fcf.slice/crio-f7d1788cbc507155f860b3041675cec343360fab6f364efe3fb1134c7518bd37 WatchSource:0}: Error finding container f7d1788cbc507155f860b3041675cec343360fab6f364efe3fb1134c7518bd37: Status 404 returned error can't find the container with id f7d1788cbc507155f860b3041675cec343360fab6f364efe3fb1134c7518bd37 Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.009685 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d482296c-b6fe-46a1-8798-696c4d06cdf0" path="/var/lib/kubelet/pods/d482296c-b6fe-46a1-8798-696c4d06cdf0/volumes" Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.010316 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f244-account-create-update-kc8sg"] Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.011176 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.012453 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fd5e-account-create-update-m58kx" event={"ID":"f82da625-890e-43b7-8dee-7ab76e375b52","Type":"ContainerStarted","Data":"e4e49c8e14724339528fd40f102fe47b4d97119994894531101083843d09ead7"} Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.012492 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fd5e-account-create-update-m58kx" event={"ID":"f82da625-890e-43b7-8dee-7ab76e375b52","Type":"ContainerStarted","Data":"6ceb54be1d397bde47a136d88a52a756ee9c63441e01fc1a1fa7cad1fd87f914"} Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.026485 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j6mch" event={"ID":"e2aacebd-28b2-43f7-afaa-966bd5bf7ade","Type":"ContainerStarted","Data":"230656781d2e7c4b1b9fc47f342546e1fd086719a5ed2672d490bb1ed7e69539"} Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.026525 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j6mch" event={"ID":"e2aacebd-28b2-43f7-afaa-966bd5bf7ade","Type":"ContainerStarted","Data":"7eb3fa960370acd110e97c605d010c4d83d9bcf85da3447335036c96f72665dc"} Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.063186 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d27lr"] Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.086094 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-fd5e-account-create-update-m58kx" podStartSLOduration=2.086077848 podStartE2EDuration="2.086077848s" podCreationTimestamp="2026-02-15 20:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:33.081316431 +0000 UTC m=+1080.947332054" watchObservedRunningTime="2026-02-15 20:34:33.086077848 +0000 UTC m=+1080.952093471" Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.273047 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-j6mch" podStartSLOduration=2.2730142620000002 podStartE2EDuration="2.273014262s" podCreationTimestamp="2026-02-15 20:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:33.125682027 +0000 UTC m=+1080.991697650" watchObservedRunningTime="2026-02-15 20:34:33.273014262 +0000 UTC m=+1081.139029885" Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.285075 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ft5v7"] Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.392482 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-p2mmq"] Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.595285 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8ea4-account-create-update-sjn94"] Feb 15 20:34:33 crc kubenswrapper[4735]: I0215 20:34:33.612380 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.033914 4735 generic.go:334] "Generic (PLEG): container finished" podID="36b69cc5-60b4-4560-9a60-27db0295a1cc" containerID="b284e77e8f04e551f96ea3329e76b0508cdfd5b5047ca426e3bb254e9860f63e" exitCode=0 Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.034020 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d27lr" event={"ID":"36b69cc5-60b4-4560-9a60-27db0295a1cc","Type":"ContainerDied","Data":"b284e77e8f04e551f96ea3329e76b0508cdfd5b5047ca426e3bb254e9860f63e"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.034216 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d27lr" event={"ID":"36b69cc5-60b4-4560-9a60-27db0295a1cc","Type":"ContainerStarted","Data":"50d3482f0489e15ddc5a1e9b7c6a7cfa0d5a287bcca2c48e40b94a3060518ef1"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.038309 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8ea4-account-create-update-sjn94" event={"ID":"b66e376f-3330-4024-a409-c6820011d354","Type":"ContainerStarted","Data":"de1b3c92adaf1449e8bbe0b59c65e47e6120fb3f52283281fb50db3bb7a8e1aa"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.038370 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8ea4-account-create-update-sjn94" event={"ID":"b66e376f-3330-4024-a409-c6820011d354","Type":"ContainerStarted","Data":"fb639f735b467199b17d852688af8c6d493d180dbc4bf5e4d12b933adb9e1459"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.039976 4735 generic.go:334] "Generic (PLEG): container finished" podID="f82da625-890e-43b7-8dee-7ab76e375b52" containerID="e4e49c8e14724339528fd40f102fe47b4d97119994894531101083843d09ead7" exitCode=0 Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.040047 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fd5e-account-create-update-m58kx" event={"ID":"f82da625-890e-43b7-8dee-7ab76e375b52","Type":"ContainerDied","Data":"e4e49c8e14724339528fd40f102fe47b4d97119994894531101083843d09ead7"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.041123 4735 generic.go:334] "Generic (PLEG): container finished" podID="b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66" containerID="4e51bbb9c8a5e7751e1502d3e29d703524a71a0e0da362fb564c6b50c832f5a9" exitCode=0 Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.041163 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-p2mmq" event={"ID":"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66","Type":"ContainerDied","Data":"4e51bbb9c8a5e7751e1502d3e29d703524a71a0e0da362fb564c6b50c832f5a9"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.041193 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-p2mmq" event={"ID":"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66","Type":"ContainerStarted","Data":"42f6468312ae390c2f11d595bed72811c968a65d124a14b997ad7049a63d3777"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.052273 4735 generic.go:334] "Generic (PLEG): container finished" podID="e2aacebd-28b2-43f7-afaa-966bd5bf7ade" containerID="230656781d2e7c4b1b9fc47f342546e1fd086719a5ed2672d490bb1ed7e69539" exitCode=0 Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.052385 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j6mch" event={"ID":"e2aacebd-28b2-43f7-afaa-966bd5bf7ade","Type":"ContainerDied","Data":"230656781d2e7c4b1b9fc47f342546e1fd086719a5ed2672d490bb1ed7e69539"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.055237 4735 generic.go:334] "Generic (PLEG): container finished" podID="d3bc62da-42b0-4f89-907a-64b5d2443fcf" containerID="a5bc5afa03916a5704a9ec8c371f7ae1f88bd238482b4f6a7ac79a273e6ec903" exitCode=0 Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.055296 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f244-account-create-update-kc8sg" event={"ID":"d3bc62da-42b0-4f89-907a-64b5d2443fcf","Type":"ContainerDied","Data":"a5bc5afa03916a5704a9ec8c371f7ae1f88bd238482b4f6a7ac79a273e6ec903"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.055317 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f244-account-create-update-kc8sg" event={"ID":"d3bc62da-42b0-4f89-907a-64b5d2443fcf","Type":"ContainerStarted","Data":"f7d1788cbc507155f860b3041675cec343360fab6f364efe3fb1134c7518bd37"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.056952 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ft5v7" event={"ID":"5c3905d6-692c-4791-84b0-58d7f6372642","Type":"ContainerStarted","Data":"681d97ffda76c0fcc4abe4ba51edd0081226e3504a81ac0df7f31cfd2944e863"} Feb 15 20:34:34 crc kubenswrapper[4735]: I0215 20:34:34.093498 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8ea4-account-create-update-sjn94" podStartSLOduration=2.093481106 podStartE2EDuration="2.093481106s" podCreationTimestamp="2026-02-15 20:34:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:34.087628779 +0000 UTC m=+1081.953644402" watchObservedRunningTime="2026-02-15 20:34:34.093481106 +0000 UTC m=+1081.959496719" Feb 15 20:34:35 crc kubenswrapper[4735]: I0215 20:34:35.068800 4735 generic.go:334] "Generic (PLEG): container finished" podID="b66e376f-3330-4024-a409-c6820011d354" containerID="de1b3c92adaf1449e8bbe0b59c65e47e6120fb3f52283281fb50db3bb7a8e1aa" exitCode=0 Feb 15 20:34:35 crc kubenswrapper[4735]: I0215 20:34:35.069488 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8ea4-account-create-update-sjn94" event={"ID":"b66e376f-3330-4024-a409-c6820011d354","Type":"ContainerDied","Data":"de1b3c92adaf1449e8bbe0b59c65e47e6120fb3f52283281fb50db3bb7a8e1aa"} Feb 15 20:34:35 crc kubenswrapper[4735]: I0215 20:34:35.384690 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:35 crc kubenswrapper[4735]: I0215 20:34:35.542004 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqphw\" (UniqueName: \"kubernetes.io/projected/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-kube-api-access-vqphw\") pod \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\" (UID: \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\") " Feb 15 20:34:35 crc kubenswrapper[4735]: I0215 20:34:35.542233 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-operator-scripts\") pod \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\" (UID: \"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66\") " Feb 15 20:34:35 crc kubenswrapper[4735]: I0215 20:34:35.542914 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66" (UID: "b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:35 crc kubenswrapper[4735]: I0215 20:34:35.547833 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-kube-api-access-vqphw" (OuterVolumeSpecName: "kube-api-access-vqphw") pod "b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66" (UID: "b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66"). InnerVolumeSpecName "kube-api-access-vqphw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:35 crc kubenswrapper[4735]: I0215 20:34:35.643996 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqphw\" (UniqueName: \"kubernetes.io/projected/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-kube-api-access-vqphw\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:35 crc kubenswrapper[4735]: I0215 20:34:35.644026 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:36 crc kubenswrapper[4735]: I0215 20:34:36.121306 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p2mmq" Feb 15 20:34:36 crc kubenswrapper[4735]: I0215 20:34:36.121682 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-p2mmq" event={"ID":"b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66","Type":"ContainerDied","Data":"42f6468312ae390c2f11d595bed72811c968a65d124a14b997ad7049a63d3777"} Feb 15 20:34:36 crc kubenswrapper[4735]: I0215 20:34:36.121727 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42f6468312ae390c2f11d595bed72811c968a65d124a14b997ad7049a63d3777" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.661933 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.668560 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.691874 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.765314 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.771489 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.801336 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bc62da-42b0-4f89-907a-64b5d2443fcf-operator-scripts\") pod \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\" (UID: \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.801395 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k22pr\" (UniqueName: \"kubernetes.io/projected/b66e376f-3330-4024-a409-c6820011d354-kube-api-access-k22pr\") pod \"b66e376f-3330-4024-a409-c6820011d354\" (UID: \"b66e376f-3330-4024-a409-c6820011d354\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.801470 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66e376f-3330-4024-a409-c6820011d354-operator-scripts\") pod \"b66e376f-3330-4024-a409-c6820011d354\" (UID: \"b66e376f-3330-4024-a409-c6820011d354\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.801491 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7qbv\" (UniqueName: \"kubernetes.io/projected/d3bc62da-42b0-4f89-907a-64b5d2443fcf-kube-api-access-j7qbv\") pod \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\" (UID: \"d3bc62da-42b0-4f89-907a-64b5d2443fcf\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.801539 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f82da625-890e-43b7-8dee-7ab76e375b52-operator-scripts\") pod \"f82da625-890e-43b7-8dee-7ab76e375b52\" (UID: \"f82da625-890e-43b7-8dee-7ab76e375b52\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.801566 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flxg4\" (UniqueName: \"kubernetes.io/projected/f82da625-890e-43b7-8dee-7ab76e375b52-kube-api-access-flxg4\") pod \"f82da625-890e-43b7-8dee-7ab76e375b52\" (UID: \"f82da625-890e-43b7-8dee-7ab76e375b52\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.803646 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f82da625-890e-43b7-8dee-7ab76e375b52-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f82da625-890e-43b7-8dee-7ab76e375b52" (UID: "f82da625-890e-43b7-8dee-7ab76e375b52"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.803641 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3bc62da-42b0-4f89-907a-64b5d2443fcf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3bc62da-42b0-4f89-907a-64b5d2443fcf" (UID: "d3bc62da-42b0-4f89-907a-64b5d2443fcf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.803681 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b66e376f-3330-4024-a409-c6820011d354-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b66e376f-3330-4024-a409-c6820011d354" (UID: "b66e376f-3330-4024-a409-c6820011d354"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.808602 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3bc62da-42b0-4f89-907a-64b5d2443fcf-kube-api-access-j7qbv" (OuterVolumeSpecName: "kube-api-access-j7qbv") pod "d3bc62da-42b0-4f89-907a-64b5d2443fcf" (UID: "d3bc62da-42b0-4f89-907a-64b5d2443fcf"). InnerVolumeSpecName "kube-api-access-j7qbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.808901 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f82da625-890e-43b7-8dee-7ab76e375b52-kube-api-access-flxg4" (OuterVolumeSpecName: "kube-api-access-flxg4") pod "f82da625-890e-43b7-8dee-7ab76e375b52" (UID: "f82da625-890e-43b7-8dee-7ab76e375b52"). InnerVolumeSpecName "kube-api-access-flxg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.816915 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b66e376f-3330-4024-a409-c6820011d354-kube-api-access-k22pr" (OuterVolumeSpecName: "kube-api-access-k22pr") pod "b66e376f-3330-4024-a409-c6820011d354" (UID: "b66e376f-3330-4024-a409-c6820011d354"). InnerVolumeSpecName "kube-api-access-k22pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.911788 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36b69cc5-60b4-4560-9a60-27db0295a1cc-operator-scripts\") pod \"36b69cc5-60b4-4560-9a60-27db0295a1cc\" (UID: \"36b69cc5-60b4-4560-9a60-27db0295a1cc\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.912009 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4lj2\" (UniqueName: \"kubernetes.io/projected/36b69cc5-60b4-4560-9a60-27db0295a1cc-kube-api-access-c4lj2\") pod \"36b69cc5-60b4-4560-9a60-27db0295a1cc\" (UID: \"36b69cc5-60b4-4560-9a60-27db0295a1cc\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.912564 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-operator-scripts\") pod \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\" (UID: \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.912624 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zhvb\" (UniqueName: \"kubernetes.io/projected/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-kube-api-access-5zhvb\") pod \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\" (UID: \"e2aacebd-28b2-43f7-afaa-966bd5bf7ade\") " Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.913437 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e2aacebd-28b2-43f7-afaa-966bd5bf7ade" (UID: "e2aacebd-28b2-43f7-afaa-966bd5bf7ade"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.913435 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36b69cc5-60b4-4560-9a60-27db0295a1cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "36b69cc5-60b4-4560-9a60-27db0295a1cc" (UID: "36b69cc5-60b4-4560-9a60-27db0295a1cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.914357 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.914384 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bc62da-42b0-4f89-907a-64b5d2443fcf-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.914405 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k22pr\" (UniqueName: \"kubernetes.io/projected/b66e376f-3330-4024-a409-c6820011d354-kube-api-access-k22pr\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.914420 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36b69cc5-60b4-4560-9a60-27db0295a1cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.914433 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66e376f-3330-4024-a409-c6820011d354-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.914446 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7qbv\" (UniqueName: \"kubernetes.io/projected/d3bc62da-42b0-4f89-907a-64b5d2443fcf-kube-api-access-j7qbv\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.914464 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f82da625-890e-43b7-8dee-7ab76e375b52-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.914476 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flxg4\" (UniqueName: \"kubernetes.io/projected/f82da625-890e-43b7-8dee-7ab76e375b52-kube-api-access-flxg4\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.920517 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-kube-api-access-5zhvb" (OuterVolumeSpecName: "kube-api-access-5zhvb") pod "e2aacebd-28b2-43f7-afaa-966bd5bf7ade" (UID: "e2aacebd-28b2-43f7-afaa-966bd5bf7ade"). InnerVolumeSpecName "kube-api-access-5zhvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:38 crc kubenswrapper[4735]: I0215 20:34:38.920670 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36b69cc5-60b4-4560-9a60-27db0295a1cc-kube-api-access-c4lj2" (OuterVolumeSpecName: "kube-api-access-c4lj2") pod "36b69cc5-60b4-4560-9a60-27db0295a1cc" (UID: "36b69cc5-60b4-4560-9a60-27db0295a1cc"). InnerVolumeSpecName "kube-api-access-c4lj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.018408 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zhvb\" (UniqueName: \"kubernetes.io/projected/e2aacebd-28b2-43f7-afaa-966bd5bf7ade-kube-api-access-5zhvb\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.018449 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4lj2\" (UniqueName: \"kubernetes.io/projected/36b69cc5-60b4-4560-9a60-27db0295a1cc-kube-api-access-c4lj2\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.141328 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fd5e-account-create-update-m58kx" event={"ID":"f82da625-890e-43b7-8dee-7ab76e375b52","Type":"ContainerDied","Data":"6ceb54be1d397bde47a136d88a52a756ee9c63441e01fc1a1fa7cad1fd87f914"} Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.141372 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ceb54be1d397bde47a136d88a52a756ee9c63441e01fc1a1fa7cad1fd87f914" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.141390 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fd5e-account-create-update-m58kx" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.143545 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8ea4-account-create-update-sjn94" event={"ID":"b66e376f-3330-4024-a409-c6820011d354","Type":"ContainerDied","Data":"fb639f735b467199b17d852688af8c6d493d180dbc4bf5e4d12b933adb9e1459"} Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.143587 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb639f735b467199b17d852688af8c6d493d180dbc4bf5e4d12b933adb9e1459" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.143647 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8ea4-account-create-update-sjn94" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.148434 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j6mch" event={"ID":"e2aacebd-28b2-43f7-afaa-966bd5bf7ade","Type":"ContainerDied","Data":"7eb3fa960370acd110e97c605d010c4d83d9bcf85da3447335036c96f72665dc"} Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.148478 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eb3fa960370acd110e97c605d010c4d83d9bcf85da3447335036c96f72665dc" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.148572 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j6mch" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.174683 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ft5v7" event={"ID":"5c3905d6-692c-4791-84b0-58d7f6372642","Type":"ContainerStarted","Data":"b646967960384495241fdd563402509ed82e8749048e7b71208e69151ee8e5af"} Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.178480 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d27lr" event={"ID":"36b69cc5-60b4-4560-9a60-27db0295a1cc","Type":"ContainerDied","Data":"50d3482f0489e15ddc5a1e9b7c6a7cfa0d5a287bcca2c48e40b94a3060518ef1"} Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.178634 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50d3482f0489e15ddc5a1e9b7c6a7cfa0d5a287bcca2c48e40b94a3060518ef1" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.178833 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d27lr" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.180895 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f244-account-create-update-kc8sg" event={"ID":"d3bc62da-42b0-4f89-907a-64b5d2443fcf","Type":"ContainerDied","Data":"f7d1788cbc507155f860b3041675cec343360fab6f364efe3fb1134c7518bd37"} Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.181051 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7d1788cbc507155f860b3041675cec343360fab6f364efe3fb1134c7518bd37" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.181249 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f244-account-create-update-kc8sg" Feb 15 20:34:39 crc kubenswrapper[4735]: I0215 20:34:39.200396 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ft5v7" podStartSLOduration=2.975998437 podStartE2EDuration="8.200378759s" podCreationTimestamp="2026-02-15 20:34:31 +0000 UTC" firstStartedPulling="2026-02-15 20:34:33.312260121 +0000 UTC m=+1081.178275744" lastFinishedPulling="2026-02-15 20:34:38.536640433 +0000 UTC m=+1086.402656066" observedRunningTime="2026-02-15 20:34:39.193833243 +0000 UTC m=+1087.059848866" watchObservedRunningTime="2026-02-15 20:34:39.200378759 +0000 UTC m=+1087.066394382" Feb 15 20:34:42 crc kubenswrapper[4735]: I0215 20:34:42.205929 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c3905d6-692c-4791-84b0-58d7f6372642" containerID="b646967960384495241fdd563402509ed82e8749048e7b71208e69151ee8e5af" exitCode=0 Feb 15 20:34:42 crc kubenswrapper[4735]: I0215 20:34:42.206163 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ft5v7" event={"ID":"5c3905d6-692c-4791-84b0-58d7f6372642","Type":"ContainerDied","Data":"b646967960384495241fdd563402509ed82e8749048e7b71208e69151ee8e5af"} Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.578680 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.699417 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-combined-ca-bundle\") pod \"5c3905d6-692c-4791-84b0-58d7f6372642\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.699491 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhrcf\" (UniqueName: \"kubernetes.io/projected/5c3905d6-692c-4791-84b0-58d7f6372642-kube-api-access-fhrcf\") pod \"5c3905d6-692c-4791-84b0-58d7f6372642\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.699629 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-config-data\") pod \"5c3905d6-692c-4791-84b0-58d7f6372642\" (UID: \"5c3905d6-692c-4791-84b0-58d7f6372642\") " Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.718117 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c3905d6-692c-4791-84b0-58d7f6372642-kube-api-access-fhrcf" (OuterVolumeSpecName: "kube-api-access-fhrcf") pod "5c3905d6-692c-4791-84b0-58d7f6372642" (UID: "5c3905d6-692c-4791-84b0-58d7f6372642"). InnerVolumeSpecName "kube-api-access-fhrcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.735173 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c3905d6-692c-4791-84b0-58d7f6372642" (UID: "5c3905d6-692c-4791-84b0-58d7f6372642"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.759876 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-config-data" (OuterVolumeSpecName: "config-data") pod "5c3905d6-692c-4791-84b0-58d7f6372642" (UID: "5c3905d6-692c-4791-84b0-58d7f6372642"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.801497 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.801532 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c3905d6-692c-4791-84b0-58d7f6372642-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:43 crc kubenswrapper[4735]: I0215 20:34:43.801546 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhrcf\" (UniqueName: \"kubernetes.io/projected/5c3905d6-692c-4791-84b0-58d7f6372642-kube-api-access-fhrcf\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.225402 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ft5v7" event={"ID":"5c3905d6-692c-4791-84b0-58d7f6372642","Type":"ContainerDied","Data":"681d97ffda76c0fcc4abe4ba51edd0081226e3504a81ac0df7f31cfd2944e863"} Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.225644 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="681d97ffda76c0fcc4abe4ba51edd0081226e3504a81ac0df7f31cfd2944e863" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.225717 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ft5v7" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.637470 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-ggkz4"] Feb 15 20:34:44 crc kubenswrapper[4735]: E0215 20:34:44.637925 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b69cc5-60b4-4560-9a60-27db0295a1cc" containerName="mariadb-database-create" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.637972 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b69cc5-60b4-4560-9a60-27db0295a1cc" containerName="mariadb-database-create" Feb 15 20:34:44 crc kubenswrapper[4735]: E0215 20:34:44.637985 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66e376f-3330-4024-a409-c6820011d354" containerName="mariadb-account-create-update" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.637991 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66e376f-3330-4024-a409-c6820011d354" containerName="mariadb-account-create-update" Feb 15 20:34:44 crc kubenswrapper[4735]: E0215 20:34:44.638003 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82da625-890e-43b7-8dee-7ab76e375b52" containerName="mariadb-account-create-update" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638009 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82da625-890e-43b7-8dee-7ab76e375b52" containerName="mariadb-account-create-update" Feb 15 20:34:44 crc kubenswrapper[4735]: E0215 20:34:44.638022 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3bc62da-42b0-4f89-907a-64b5d2443fcf" containerName="mariadb-account-create-update" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638028 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3bc62da-42b0-4f89-907a-64b5d2443fcf" containerName="mariadb-account-create-update" Feb 15 20:34:44 crc kubenswrapper[4735]: E0215 20:34:44.638049 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66" containerName="mariadb-database-create" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638071 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66" containerName="mariadb-database-create" Feb 15 20:34:44 crc kubenswrapper[4735]: E0215 20:34:44.638083 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2aacebd-28b2-43f7-afaa-966bd5bf7ade" containerName="mariadb-database-create" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638089 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2aacebd-28b2-43f7-afaa-966bd5bf7ade" containerName="mariadb-database-create" Feb 15 20:34:44 crc kubenswrapper[4735]: E0215 20:34:44.638101 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c3905d6-692c-4791-84b0-58d7f6372642" containerName="keystone-db-sync" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638107 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c3905d6-692c-4791-84b0-58d7f6372642" containerName="keystone-db-sync" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638266 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3bc62da-42b0-4f89-907a-64b5d2443fcf" containerName="mariadb-account-create-update" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638277 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2aacebd-28b2-43f7-afaa-966bd5bf7ade" containerName="mariadb-database-create" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638287 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f82da625-890e-43b7-8dee-7ab76e375b52" containerName="mariadb-account-create-update" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638297 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66" containerName="mariadb-database-create" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638308 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c3905d6-692c-4791-84b0-58d7f6372642" containerName="keystone-db-sync" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638324 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="36b69cc5-60b4-4560-9a60-27db0295a1cc" containerName="mariadb-database-create" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.638332 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b66e376f-3330-4024-a409-c6820011d354" containerName="mariadb-account-create-update" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.640504 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.672000 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-622mp"] Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.673275 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.679432 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.679659 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.679988 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.686046 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-ggkz4"] Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.688251 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dljv4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.688485 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.700530 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-622mp"] Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.732933 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.733005 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89749\" (UniqueName: \"kubernetes.io/projected/2d26a448-2656-4f80-b5b0-1b51d7c19c21-kube-api-access-89749\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.733059 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.733084 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.733134 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-config\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.733159 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.834689 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-config\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835125 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835223 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-config-data\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835303 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-combined-ca-bundle\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835389 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835456 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-scripts\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835518 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-fernet-keys\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835590 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89749\" (UniqueName: \"kubernetes.io/projected/2d26a448-2656-4f80-b5b0-1b51d7c19c21-kube-api-access-89749\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835663 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-credential-keys\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835760 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835831 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.835899 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krpng\" (UniqueName: \"kubernetes.io/projected/09c547be-bb78-4d22-adcf-3643dd828a78-kube-api-access-krpng\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.836731 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-config\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.837395 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.837997 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.838831 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.839424 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.864740 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-6htdh"] Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.865925 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.872540 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.872714 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.872822 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-76wr6" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.920694 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89749\" (UniqueName: \"kubernetes.io/projected/2d26a448-2656-4f80-b5b0-1b51d7c19c21-kube-api-access-89749\") pod \"dnsmasq-dns-bbf5cc879-ggkz4\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.936994 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krpng\" (UniqueName: \"kubernetes.io/projected/09c547be-bb78-4d22-adcf-3643dd828a78-kube-api-access-krpng\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.937104 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-config-data\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.937125 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-combined-ca-bundle\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.937162 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-scripts\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.937181 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-fernet-keys\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.937203 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-credential-keys\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.940650 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-combined-ca-bundle\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.947421 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-scripts\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.948382 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-fernet-keys\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.949936 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6htdh"] Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.950315 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-credential-keys\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.964847 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-config-data\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.966381 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.968372 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-hk7jh"] Feb 15 20:34:44 crc kubenswrapper[4735]: I0215 20:34:44.969319 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.003522 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.003736 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-w7s2g" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.003815 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.014501 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krpng\" (UniqueName: \"kubernetes.io/projected/09c547be-bb78-4d22-adcf-3643dd828a78-kube-api-access-krpng\") pod \"keystone-bootstrap-622mp\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.040165 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74f7d59545-hbprm"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.042027 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.046687 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-7dw5m" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.046927 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.047061 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.047181 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.055099 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-config-data\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.055168 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-db-sync-config-data\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.055261 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec4fe5e3-ca92-4814-8921-62117a037281-etc-machine-id\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.055284 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-scripts\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.055306 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxvj5\" (UniqueName: \"kubernetes.io/projected/ec4fe5e3-ca92-4814-8921-62117a037281-kube-api-access-kxvj5\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.055329 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-combined-ca-bundle\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.106511 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hk7jh"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.141426 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74f7d59545-hbprm"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.157133 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-config-data\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.157367 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d787c\" (UniqueName: \"kubernetes.io/projected/1be12817-7c08-4d66-938d-98ce9f79e4fd-kube-api-access-d787c\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.157460 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-db-sync-config-data\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.157538 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-config-data\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.157608 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1be12817-7c08-4d66-938d-98ce9f79e4fd-logs\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.157714 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-config\") pod \"neutron-db-sync-hk7jh\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.157782 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plkhg\" (UniqueName: \"kubernetes.io/projected/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-kube-api-access-plkhg\") pod \"neutron-db-sync-hk7jh\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.157858 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-scripts\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.157930 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec4fe5e3-ca92-4814-8921-62117a037281-etc-machine-id\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.158014 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-scripts\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.158101 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxvj5\" (UniqueName: \"kubernetes.io/projected/ec4fe5e3-ca92-4814-8921-62117a037281-kube-api-access-kxvj5\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.158173 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-combined-ca-bundle\") pod \"neutron-db-sync-hk7jh\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.158240 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-combined-ca-bundle\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.158326 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1be12817-7c08-4d66-938d-98ce9f79e4fd-horizon-secret-key\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.158499 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec4fe5e3-ca92-4814-8921-62117a037281-etc-machine-id\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.164458 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-combined-ca-bundle\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.165969 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-config-data\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.176554 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-scripts\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.196431 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-db-sync-config-data\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.214005 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-pss7t"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.215572 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.218258 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxvj5\" (UniqueName: \"kubernetes.io/projected/ec4fe5e3-ca92-4814-8921-62117a037281-kube-api-access-kxvj5\") pod \"cinder-db-sync-6htdh\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.222635 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-tblss" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.227168 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-ggkz4"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.240435 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pss7t"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.223817 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.224366 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.263621 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-combined-ca-bundle\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264095 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-config-data\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264125 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-config-data\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264145 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1be12817-7c08-4d66-938d-98ce9f79e4fd-logs\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264181 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-config\") pod \"neutron-db-sync-hk7jh\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264202 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plkhg\" (UniqueName: \"kubernetes.io/projected/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-kube-api-access-plkhg\") pod \"neutron-db-sync-hk7jh\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264231 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-scripts\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264259 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvbgv\" (UniqueName: \"kubernetes.io/projected/327449fc-a14d-4594-b8a8-c758aeebec34-kube-api-access-zvbgv\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264284 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327449fc-a14d-4594-b8a8-c758aeebec34-logs\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264312 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-combined-ca-bundle\") pod \"neutron-db-sync-hk7jh\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264347 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1be12817-7c08-4d66-938d-98ce9f79e4fd-horizon-secret-key\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264382 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-scripts\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.264413 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d787c\" (UniqueName: \"kubernetes.io/projected/1be12817-7c08-4d66-938d-98ce9f79e4fd-kube-api-access-d787c\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.266247 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-config-data\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.266734 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-scripts\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.267052 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1be12817-7c08-4d66-938d-98ce9f79e4fd-logs\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.270741 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-combined-ca-bundle\") pod \"neutron-db-sync-hk7jh\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.276150 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-config\") pod \"neutron-db-sync-hk7jh\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.276358 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1be12817-7c08-4d66-938d-98ce9f79e4fd-horizon-secret-key\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.288018 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ph5qb"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.290097 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.298030 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74dd7854c7-t2t6r"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.301238 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.301666 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-b9t7b" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.301922 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.301921 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plkhg\" (UniqueName: \"kubernetes.io/projected/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-kube-api-access-plkhg\") pod \"neutron-db-sync-hk7jh\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.307373 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-622mp" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.316585 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ph5qb"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369336 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-scripts\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369401 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/faf23e04-da47-4121-9e55-d003d8d6dcfe-horizon-secret-key\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369440 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-combined-ca-bundle\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369501 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzrk6\" (UniqueName: \"kubernetes.io/projected/76a8192d-7bcd-4f4a-b798-907d67b3d21c-kube-api-access-kzrk6\") pod \"barbican-db-sync-ph5qb\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369545 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-config-data\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369571 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-config-data\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369652 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfbp4\" (UniqueName: \"kubernetes.io/projected/faf23e04-da47-4121-9e55-d003d8d6dcfe-kube-api-access-mfbp4\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369678 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-db-sync-config-data\") pod \"barbican-db-sync-ph5qb\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369700 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-combined-ca-bundle\") pod \"barbican-db-sync-ph5qb\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369729 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvbgv\" (UniqueName: \"kubernetes.io/projected/327449fc-a14d-4594-b8a8-c758aeebec34-kube-api-access-zvbgv\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369773 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327449fc-a14d-4594-b8a8-c758aeebec34-logs\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369811 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faf23e04-da47-4121-9e55-d003d8d6dcfe-logs\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.369877 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-scripts\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.381021 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327449fc-a14d-4594-b8a8-c758aeebec34-logs\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.409876 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d787c\" (UniqueName: \"kubernetes.io/projected/1be12817-7c08-4d66-938d-98ce9f79e4fd-kube-api-access-d787c\") pod \"horizon-74f7d59545-hbprm\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.410145 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-config-data\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.410812 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-combined-ca-bundle\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.410901 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.429881 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.433762 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-s945p"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.439201 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.450149 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-scripts\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.463688 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvbgv\" (UniqueName: \"kubernetes.io/projected/327449fc-a14d-4594-b8a8-c758aeebec34-kube-api-access-zvbgv\") pod \"placement-db-sync-pss7t\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.479445 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74dd7854c7-t2t6r"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.481269 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6htdh" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.518135 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.518201 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.518260 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfbp4\" (UniqueName: \"kubernetes.io/projected/faf23e04-da47-4121-9e55-d003d8d6dcfe-kube-api-access-mfbp4\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.518285 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-db-sync-config-data\") pod \"barbican-db-sync-ph5qb\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.518305 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-combined-ca-bundle\") pod \"barbican-db-sync-ph5qb\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.518384 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faf23e04-da47-4121-9e55-d003d8d6dcfe-logs\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.518415 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.518453 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-config\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.521212 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-scripts\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.521255 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/faf23e04-da47-4121-9e55-d003d8d6dcfe-horizon-secret-key\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.521276 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.521312 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbrsc\" (UniqueName: \"kubernetes.io/projected/da52e1cb-1436-4b18-a2b6-5032d1018b0b-kube-api-access-fbrsc\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.521348 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzrk6\" (UniqueName: \"kubernetes.io/projected/76a8192d-7bcd-4f4a-b798-907d67b3d21c-kube-api-access-kzrk6\") pod \"barbican-db-sync-ph5qb\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.521404 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-config-data\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.522768 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-scripts\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.524936 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-config-data\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.531661 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faf23e04-da47-4121-9e55-d003d8d6dcfe-logs\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.546341 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.546835 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pss7t" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.547737 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.551612 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-combined-ca-bundle\") pod \"barbican-db-sync-ph5qb\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.561833 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/faf23e04-da47-4121-9e55-d003d8d6dcfe-horizon-secret-key\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.563387 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-db-sync-config-data\") pod \"barbican-db-sync-ph5qb\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.579303 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-s945p"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.594060 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.594667 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.594785 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.594888 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kfjgg" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.617008 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.628708 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.628753 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbrsc\" (UniqueName: \"kubernetes.io/projected/da52e1cb-1436-4b18-a2b6-5032d1018b0b-kube-api-access-fbrsc\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.628802 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.628824 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.628867 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.628887 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-config\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.637432 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.642806 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.643357 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.645708 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.648638 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-config\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.664551 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfbp4\" (UniqueName: \"kubernetes.io/projected/faf23e04-da47-4121-9e55-d003d8d6dcfe-kube-api-access-mfbp4\") pod \"horizon-74dd7854c7-t2t6r\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.698426 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzrk6\" (UniqueName: \"kubernetes.io/projected/76a8192d-7bcd-4f4a-b798-907d67b3d21c-kube-api-access-kzrk6\") pod \"barbican-db-sync-ph5qb\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.698803 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.698803 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbrsc\" (UniqueName: \"kubernetes.io/projected/da52e1cb-1436-4b18-a2b6-5032d1018b0b-kube-api-access-fbrsc\") pod \"dnsmasq-dns-56df8fb6b7-s945p\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.705460 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.707595 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.720434 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.720628 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.752690 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.752974 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.753001 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.753023 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmbfh\" (UniqueName: \"kubernetes.io/projected/a0f5ac85-f76f-4aee-a16b-df6e2e549276-kube-api-access-tmbfh\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.753050 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.753066 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-logs\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.753099 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.753119 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.757083 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.801130 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.803786 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.805222 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.814305 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.824338 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.825398 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.859523 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.859576 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.859620 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmbfh\" (UniqueName: \"kubernetes.io/projected/a0f5ac85-f76f-4aee-a16b-df6e2e549276-kube-api-access-tmbfh\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.859641 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-log-httpd\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.859922 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.859958 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-logs\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.859991 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62lg9\" (UniqueName: \"kubernetes.io/projected/5d7a00e9-2271-49b1-a74f-51f5f668995a-kube-api-access-62lg9\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.860106 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-run-httpd\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.860684 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-logs\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.862317 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.862395 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.862432 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.862465 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-scripts\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.862546 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.862644 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-config-data\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.862705 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.863184 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.863544 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.867642 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.874379 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.882667 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.883173 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.904541 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmbfh\" (UniqueName: \"kubernetes.io/projected/a0f5ac85-f76f-4aee-a16b-df6e2e549276-kube-api-access-tmbfh\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.934353 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.965917 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-run-httpd\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.965983 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966011 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-scripts\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966040 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966063 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966114 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966130 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-config-data\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966148 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thxp6\" (UniqueName: \"kubernetes.io/projected/d033877b-3f4d-4d60-85da-6b029ad9856c-kube-api-access-thxp6\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966182 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966218 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-logs\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966233 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966252 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966273 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-log-httpd\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966298 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966322 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62lg9\" (UniqueName: \"kubernetes.io/projected/5d7a00e9-2271-49b1-a74f-51f5f668995a-kube-api-access-62lg9\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.966979 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-run-httpd\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.971249 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-scripts\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.976437 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-log-httpd\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.981504 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " pod="openstack/glance-default-external-api-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.982315 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.983784 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:45 crc kubenswrapper[4735]: I0215 20:34:45.983975 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-config-data\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.011470 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62lg9\" (UniqueName: \"kubernetes.io/projected/5d7a00e9-2271-49b1-a74f-51f5f668995a-kube-api-access-62lg9\") pod \"ceilometer-0\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " pod="openstack/ceilometer-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.090979 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.095357 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.103446 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.109553 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.109671 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.109842 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.109872 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thxp6\" (UniqueName: \"kubernetes.io/projected/d033877b-3f4d-4d60-85da-6b029ad9856c-kube-api-access-thxp6\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.109971 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.110003 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-logs\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.110053 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.110109 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.113452 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-logs\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.141923 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.142578 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.155182 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.155600 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.171738 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.177354 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thxp6\" (UniqueName: \"kubernetes.io/projected/d033877b-3f4d-4d60-85da-6b029ad9856c-kube-api-access-thxp6\") pod \"glance-default-internal-api-0\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.191429 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.223726 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-ggkz4"] Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.341691 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-622mp"] Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.457585 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.799769 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pss7t"] Feb 15 20:34:46 crc kubenswrapper[4735]: I0215 20:34:46.853594 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6htdh"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.110015 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hk7jh"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.304336 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-622mp" event={"ID":"09c547be-bb78-4d22-adcf-3643dd828a78","Type":"ContainerStarted","Data":"e937066142a5010c8285f2c4f18da54ca49cea00b016e120327ee7c2d11c3bfd"} Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.304546 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-622mp" event={"ID":"09c547be-bb78-4d22-adcf-3643dd828a78","Type":"ContainerStarted","Data":"5eb9e31a07e40f6abd808d0283e62a2d8822aed876f38c705deb46c560377996"} Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.307692 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6htdh" event={"ID":"ec4fe5e3-ca92-4814-8921-62117a037281","Type":"ContainerStarted","Data":"913e4bd190dbe2ba42c28044f10622cd9aa46d4ddf427984c1218b1e14c4468f"} Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.314607 4735 generic.go:334] "Generic (PLEG): container finished" podID="2d26a448-2656-4f80-b5b0-1b51d7c19c21" containerID="0a6cb7c45613fc3dfb9c71d528cda902e54bf16ba67e014bd2ff3b5bde8ddd84" exitCode=0 Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.314689 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" event={"ID":"2d26a448-2656-4f80-b5b0-1b51d7c19c21","Type":"ContainerDied","Data":"0a6cb7c45613fc3dfb9c71d528cda902e54bf16ba67e014bd2ff3b5bde8ddd84"} Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.314722 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" event={"ID":"2d26a448-2656-4f80-b5b0-1b51d7c19c21","Type":"ContainerStarted","Data":"af90e8c4f75e661b6f55beba9172d5cc10ae8d6df74da8b0ddef207a8991bc6d"} Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.326250 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pss7t" event={"ID":"327449fc-a14d-4594-b8a8-c758aeebec34","Type":"ContainerStarted","Data":"10ef2577c6b14f5922857b374e032dc0f1aa3b53d06fc744163ac5e5887b9b14"} Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.327816 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hk7jh" event={"ID":"6ae5b3cf-b0c9-4d46-ac26-491df4077db9","Type":"ContainerStarted","Data":"b62a09caa198afd75e115b735a1559c1d01429c5c7a5e616d2e3cef5d47a7a9d"} Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.336278 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-622mp" podStartSLOduration=3.336252017 podStartE2EDuration="3.336252017s" podCreationTimestamp="2026-02-15 20:34:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:47.321702175 +0000 UTC m=+1095.187717798" watchObservedRunningTime="2026-02-15 20:34:47.336252017 +0000 UTC m=+1095.202267640" Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.466752 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74f7d59545-hbprm"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.502739 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-s945p"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.516193 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ph5qb"] Feb 15 20:34:47 crc kubenswrapper[4735]: W0215 20:34:47.529058 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76a8192d_7bcd_4f4a_b798_907d67b3d21c.slice/crio-02617330b3775ed02a8aa95ad89b3c2f5d4d45f1e9a94e6337caf90f424865fd WatchSource:0}: Error finding container 02617330b3775ed02a8aa95ad89b3c2f5d4d45f1e9a94e6337caf90f424865fd: Status 404 returned error can't find the container with id 02617330b3775ed02a8aa95ad89b3c2f5d4d45f1e9a94e6337caf90f424865fd Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.538867 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.568840 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74dd7854c7-t2t6r"] Feb 15 20:34:47 crc kubenswrapper[4735]: W0215 20:34:47.584838 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaf23e04_da47_4121_9e55_d003d8d6dcfe.slice/crio-d861c6c2fb63c9717fc81d6f9f73ebc1c1006699b23adde1d101b4d0bb28ba8f WatchSource:0}: Error finding container d861c6c2fb63c9717fc81d6f9f73ebc1c1006699b23adde1d101b4d0bb28ba8f: Status 404 returned error can't find the container with id d861c6c2fb63c9717fc81d6f9f73ebc1c1006699b23adde1d101b4d0bb28ba8f Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.656100 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.838486 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.900112 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.934903 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74f7d59545-hbprm"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.988268 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6898675dc5-fh44t"] Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.991361 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:47 crc kubenswrapper[4735]: I0215 20:34:47.999910 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6898675dc5-fh44t"] Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.059001 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.127088 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-config-data\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.127142 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-logs\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.127223 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-horizon-secret-key\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.127314 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-scripts\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.127341 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxmf6\" (UniqueName: \"kubernetes.io/projected/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-kube-api-access-mxmf6\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.131334 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.228588 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-scripts\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.228630 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxmf6\" (UniqueName: \"kubernetes.io/projected/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-kube-api-access-mxmf6\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.228677 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-config-data\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.228701 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-logs\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.228746 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-horizon-secret-key\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.229412 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-scripts\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.230365 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-config-data\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.230821 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-logs\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.252391 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.253808 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-horizon-secret-key\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.261600 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxmf6\" (UniqueName: \"kubernetes.io/projected/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-kube-api-access-mxmf6\") pod \"horizon-6898675dc5-fh44t\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.332347 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89749\" (UniqueName: \"kubernetes.io/projected/2d26a448-2656-4f80-b5b0-1b51d7c19c21-kube-api-access-89749\") pod \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.332385 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-config\") pod \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.332410 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-svc\") pod \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.332433 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-nb\") pod \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.332473 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-sb\") pod \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.332571 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-swift-storage-0\") pod \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\" (UID: \"2d26a448-2656-4f80-b5b0-1b51d7c19c21\") " Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.340264 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d26a448-2656-4f80-b5b0-1b51d7c19c21-kube-api-access-89749" (OuterVolumeSpecName: "kube-api-access-89749") pod "2d26a448-2656-4f80-b5b0-1b51d7c19c21" (UID: "2d26a448-2656-4f80-b5b0-1b51d7c19c21"). InnerVolumeSpecName "kube-api-access-89749". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.388698 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2d26a448-2656-4f80-b5b0-1b51d7c19c21" (UID: "2d26a448-2656-4f80-b5b0-1b51d7c19c21"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.391293 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74dd7854c7-t2t6r" event={"ID":"faf23e04-da47-4121-9e55-d003d8d6dcfe","Type":"ContainerStarted","Data":"d861c6c2fb63c9717fc81d6f9f73ebc1c1006699b23adde1d101b4d0bb28ba8f"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.396292 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d26a448-2656-4f80-b5b0-1b51d7c19c21" (UID: "2d26a448-2656-4f80-b5b0-1b51d7c19c21"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.406173 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2d26a448-2656-4f80-b5b0-1b51d7c19c21" (UID: "2d26a448-2656-4f80-b5b0-1b51d7c19c21"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.414768 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2d26a448-2656-4f80-b5b0-1b51d7c19c21" (UID: "2d26a448-2656-4f80-b5b0-1b51d7c19c21"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.430135 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" event={"ID":"2d26a448-2656-4f80-b5b0-1b51d7c19c21","Type":"ContainerDied","Data":"af90e8c4f75e661b6f55beba9172d5cc10ae8d6df74da8b0ddef207a8991bc6d"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.430162 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-ggkz4" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.430200 4735 scope.go:117] "RemoveContainer" containerID="0a6cb7c45613fc3dfb9c71d528cda902e54bf16ba67e014bd2ff3b5bde8ddd84" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.434126 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.434153 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89749\" (UniqueName: \"kubernetes.io/projected/2d26a448-2656-4f80-b5b0-1b51d7c19c21-kube-api-access-89749\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.434164 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.434172 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.434181 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.435234 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-config" (OuterVolumeSpecName: "config") pod "2d26a448-2656-4f80-b5b0-1b51d7c19c21" (UID: "2d26a448-2656-4f80-b5b0-1b51d7c19c21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.439179 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ph5qb" event={"ID":"76a8192d-7bcd-4f4a-b798-907d67b3d21c","Type":"ContainerStarted","Data":"02617330b3775ed02a8aa95ad89b3c2f5d4d45f1e9a94e6337caf90f424865fd"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.442661 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" event={"ID":"da52e1cb-1436-4b18-a2b6-5032d1018b0b","Type":"ContainerStarted","Data":"355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.442686 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" event={"ID":"da52e1cb-1436-4b18-a2b6-5032d1018b0b","Type":"ContainerStarted","Data":"c695d6f282c9d598349eff69df6e0b8437be71602618da1fb4b65040ecf48136"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.453200 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0f5ac85-f76f-4aee-a16b-df6e2e549276","Type":"ContainerStarted","Data":"a4066efa518fb6129f7562c35856c6cd73911e46073d573429419f55d76857be"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.473748 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d033877b-3f4d-4d60-85da-6b029ad9856c","Type":"ContainerStarted","Data":"e638b7b6e25f2c83739fd6c7b3abe0b1cb91e3cc4aa18a26bec277c1203b4e5a"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.484600 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hk7jh" event={"ID":"6ae5b3cf-b0c9-4d46-ac26-491df4077db9","Type":"ContainerStarted","Data":"c2bf42173a97f9145aff4eb6ccff1e5f696023d0eaec1c7d82f9d307e916cc36"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.490152 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f7d59545-hbprm" event={"ID":"1be12817-7c08-4d66-938d-98ce9f79e4fd","Type":"ContainerStarted","Data":"6a47f18243090f01153e07f9b9083ee47fa55dc44e5b50014bdc718407908cc8"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.495816 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d7a00e9-2271-49b1-a74f-51f5f668995a","Type":"ContainerStarted","Data":"abe86075d510854b77bbb3615d54faf42689095565497865f59422016bdcbb4e"} Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.508416 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.516310 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-hk7jh" podStartSLOduration=4.516290612 podStartE2EDuration="4.516290612s" podCreationTimestamp="2026-02-15 20:34:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:48.507879155 +0000 UTC m=+1096.373894778" watchObservedRunningTime="2026-02-15 20:34:48.516290612 +0000 UTC m=+1096.382306235" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.535900 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d26a448-2656-4f80-b5b0-1b51d7c19c21-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.880819 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-ggkz4"] Feb 15 20:34:48 crc kubenswrapper[4735]: I0215 20:34:48.955064 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-ggkz4"] Feb 15 20:34:49 crc kubenswrapper[4735]: I0215 20:34:49.235780 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6898675dc5-fh44t"] Feb 15 20:34:49 crc kubenswrapper[4735]: I0215 20:34:49.577232 4735 generic.go:334] "Generic (PLEG): container finished" podID="da52e1cb-1436-4b18-a2b6-5032d1018b0b" containerID="355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b" exitCode=0 Feb 15 20:34:49 crc kubenswrapper[4735]: I0215 20:34:49.577692 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" event={"ID":"da52e1cb-1436-4b18-a2b6-5032d1018b0b","Type":"ContainerDied","Data":"355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b"} Feb 15 20:34:49 crc kubenswrapper[4735]: I0215 20:34:49.577751 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" event={"ID":"da52e1cb-1436-4b18-a2b6-5032d1018b0b","Type":"ContainerStarted","Data":"046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12"} Feb 15 20:34:49 crc kubenswrapper[4735]: I0215 20:34:49.578386 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:49 crc kubenswrapper[4735]: I0215 20:34:49.599010 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" podStartSLOduration=4.598994351 podStartE2EDuration="4.598994351s" podCreationTimestamp="2026-02-15 20:34:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:49.596665328 +0000 UTC m=+1097.462680951" watchObservedRunningTime="2026-02-15 20:34:49.598994351 +0000 UTC m=+1097.465009974" Feb 15 20:34:49 crc kubenswrapper[4735]: I0215 20:34:49.621338 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0f5ac85-f76f-4aee-a16b-df6e2e549276","Type":"ContainerStarted","Data":"e0ddeb09f9aca4f4a828cb390e5fc2c69d35b9fb9ffefe19b41667129797535c"} Feb 15 20:34:49 crc kubenswrapper[4735]: I0215 20:34:49.628493 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d033877b-3f4d-4d60-85da-6b029ad9856c","Type":"ContainerStarted","Data":"14f6a3590b5e07b06907bf5f440089060cab01534838fd703d5bf1f926bea989"} Feb 15 20:34:49 crc kubenswrapper[4735]: I0215 20:34:49.636079 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6898675dc5-fh44t" event={"ID":"a6f89bfa-6fb3-40d0-9741-cfddddb9a685","Type":"ContainerStarted","Data":"cacdff0549f62ad06f144926de14a801428227fe5e9b143386038fe067d4ba90"} Feb 15 20:34:50 crc kubenswrapper[4735]: I0215 20:34:50.901934 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d26a448-2656-4f80-b5b0-1b51d7c19c21" path="/var/lib/kubelet/pods/2d26a448-2656-4f80-b5b0-1b51d7c19c21/volumes" Feb 15 20:34:51 crc kubenswrapper[4735]: I0215 20:34:51.679296 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d033877b-3f4d-4d60-85da-6b029ad9856c","Type":"ContainerStarted","Data":"5da788e5749d2834c908ea6d4621ec0d73cec4b3f1e87d736d597e3647d6f8a4"} Feb 15 20:34:51 crc kubenswrapper[4735]: I0215 20:34:51.680717 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerName="glance-log" containerID="cri-o://14f6a3590b5e07b06907bf5f440089060cab01534838fd703d5bf1f926bea989" gracePeriod=30 Feb 15 20:34:51 crc kubenswrapper[4735]: I0215 20:34:51.680875 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerName="glance-httpd" containerID="cri-o://5da788e5749d2834c908ea6d4621ec0d73cec4b3f1e87d736d597e3647d6f8a4" gracePeriod=30 Feb 15 20:34:51 crc kubenswrapper[4735]: I0215 20:34:51.691848 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0f5ac85-f76f-4aee-a16b-df6e2e549276","Type":"ContainerStarted","Data":"98a6d1826e7109483ece50a3320347909ca3a08afe1fb8c7248a2b48ac3087cf"} Feb 15 20:34:51 crc kubenswrapper[4735]: I0215 20:34:51.692157 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerName="glance-log" containerID="cri-o://e0ddeb09f9aca4f4a828cb390e5fc2c69d35b9fb9ffefe19b41667129797535c" gracePeriod=30 Feb 15 20:34:51 crc kubenswrapper[4735]: I0215 20:34:51.692268 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerName="glance-httpd" containerID="cri-o://98a6d1826e7109483ece50a3320347909ca3a08afe1fb8c7248a2b48ac3087cf" gracePeriod=30 Feb 15 20:34:51 crc kubenswrapper[4735]: I0215 20:34:51.737912 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.737893193 podStartE2EDuration="6.737893193s" podCreationTimestamp="2026-02-15 20:34:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:51.706561188 +0000 UTC m=+1099.572576811" watchObservedRunningTime="2026-02-15 20:34:51.737893193 +0000 UTC m=+1099.603908816" Feb 15 20:34:51 crc kubenswrapper[4735]: I0215 20:34:51.743803 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.743794923 podStartE2EDuration="6.743794923s" podCreationTimestamp="2026-02-15 20:34:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:34:51.740820562 +0000 UTC m=+1099.606836195" watchObservedRunningTime="2026-02-15 20:34:51.743794923 +0000 UTC m=+1099.609810546" Feb 15 20:34:52 crc kubenswrapper[4735]: I0215 20:34:52.702367 4735 generic.go:334] "Generic (PLEG): container finished" podID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerID="98a6d1826e7109483ece50a3320347909ca3a08afe1fb8c7248a2b48ac3087cf" exitCode=0 Feb 15 20:34:52 crc kubenswrapper[4735]: I0215 20:34:52.702685 4735 generic.go:334] "Generic (PLEG): container finished" podID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerID="e0ddeb09f9aca4f4a828cb390e5fc2c69d35b9fb9ffefe19b41667129797535c" exitCode=143 Feb 15 20:34:52 crc kubenswrapper[4735]: I0215 20:34:52.702436 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0f5ac85-f76f-4aee-a16b-df6e2e549276","Type":"ContainerDied","Data":"98a6d1826e7109483ece50a3320347909ca3a08afe1fb8c7248a2b48ac3087cf"} Feb 15 20:34:52 crc kubenswrapper[4735]: I0215 20:34:52.702727 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0f5ac85-f76f-4aee-a16b-df6e2e549276","Type":"ContainerDied","Data":"e0ddeb09f9aca4f4a828cb390e5fc2c69d35b9fb9ffefe19b41667129797535c"} Feb 15 20:34:52 crc kubenswrapper[4735]: I0215 20:34:52.706064 4735 generic.go:334] "Generic (PLEG): container finished" podID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerID="5da788e5749d2834c908ea6d4621ec0d73cec4b3f1e87d736d597e3647d6f8a4" exitCode=0 Feb 15 20:34:52 crc kubenswrapper[4735]: I0215 20:34:52.706087 4735 generic.go:334] "Generic (PLEG): container finished" podID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerID="14f6a3590b5e07b06907bf5f440089060cab01534838fd703d5bf1f926bea989" exitCode=143 Feb 15 20:34:52 crc kubenswrapper[4735]: I0215 20:34:52.706102 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d033877b-3f4d-4d60-85da-6b029ad9856c","Type":"ContainerDied","Data":"5da788e5749d2834c908ea6d4621ec0d73cec4b3f1e87d736d597e3647d6f8a4"} Feb 15 20:34:52 crc kubenswrapper[4735]: I0215 20:34:52.706120 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d033877b-3f4d-4d60-85da-6b029ad9856c","Type":"ContainerDied","Data":"14f6a3590b5e07b06907bf5f440089060cab01534838fd703d5bf1f926bea989"} Feb 15 20:34:53 crc kubenswrapper[4735]: I0215 20:34:53.737619 4735 generic.go:334] "Generic (PLEG): container finished" podID="09c547be-bb78-4d22-adcf-3643dd828a78" containerID="e937066142a5010c8285f2c4f18da54ca49cea00b016e120327ee7c2d11c3bfd" exitCode=0 Feb 15 20:34:53 crc kubenswrapper[4735]: I0215 20:34:53.737660 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-622mp" event={"ID":"09c547be-bb78-4d22-adcf-3643dd828a78","Type":"ContainerDied","Data":"e937066142a5010c8285f2c4f18da54ca49cea00b016e120327ee7c2d11c3bfd"} Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.398194 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74dd7854c7-t2t6r"] Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.439126 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5d466d479d-wdxj6"] Feb 15 20:34:54 crc kubenswrapper[4735]: E0215 20:34:54.439502 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d26a448-2656-4f80-b5b0-1b51d7c19c21" containerName="init" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.439518 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d26a448-2656-4f80-b5b0-1b51d7c19c21" containerName="init" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.439667 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d26a448-2656-4f80-b5b0-1b51d7c19c21" containerName="init" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.440513 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.443116 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.451242 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d466d479d-wdxj6"] Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.492400 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6898675dc5-fh44t"] Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.519905 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c647fbf5b-7zs9r"] Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.521260 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.553760 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c647fbf5b-7zs9r"] Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588022 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-scripts\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588075 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-logs\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588097 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-config-data\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588179 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/970d5293-b621-45e1-a1e4-8cc176c9a148-config-data\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588210 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-secret-key\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588245 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdsm4\" (UniqueName: \"kubernetes.io/projected/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-kube-api-access-xdsm4\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588293 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/970d5293-b621-45e1-a1e4-8cc176c9a148-horizon-secret-key\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588314 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/970d5293-b621-45e1-a1e4-8cc176c9a148-scripts\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588350 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4wvc\" (UniqueName: \"kubernetes.io/projected/970d5293-b621-45e1-a1e4-8cc176c9a148-kube-api-access-l4wvc\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588364 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-tls-certs\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588386 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/970d5293-b621-45e1-a1e4-8cc176c9a148-logs\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588410 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-combined-ca-bundle\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588426 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970d5293-b621-45e1-a1e4-8cc176c9a148-combined-ca-bundle\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.588440 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/970d5293-b621-45e1-a1e4-8cc176c9a148-horizon-tls-certs\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690586 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4wvc\" (UniqueName: \"kubernetes.io/projected/970d5293-b621-45e1-a1e4-8cc176c9a148-kube-api-access-l4wvc\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690628 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-tls-certs\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690688 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/970d5293-b621-45e1-a1e4-8cc176c9a148-logs\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690713 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-combined-ca-bundle\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690744 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970d5293-b621-45e1-a1e4-8cc176c9a148-combined-ca-bundle\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690757 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/970d5293-b621-45e1-a1e4-8cc176c9a148-horizon-tls-certs\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690786 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-scripts\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690812 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-logs\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690828 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-config-data\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690861 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/970d5293-b621-45e1-a1e4-8cc176c9a148-config-data\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690891 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-secret-key\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690911 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdsm4\" (UniqueName: \"kubernetes.io/projected/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-kube-api-access-xdsm4\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.690965 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/970d5293-b621-45e1-a1e4-8cc176c9a148-horizon-secret-key\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.691101 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/970d5293-b621-45e1-a1e4-8cc176c9a148-logs\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.691316 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-logs\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.691815 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-scripts\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.692095 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/970d5293-b621-45e1-a1e4-8cc176c9a148-scripts\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.692668 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/970d5293-b621-45e1-a1e4-8cc176c9a148-scripts\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.693059 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-config-data\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.693890 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/970d5293-b621-45e1-a1e4-8cc176c9a148-config-data\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.700395 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/970d5293-b621-45e1-a1e4-8cc176c9a148-horizon-tls-certs\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.700503 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-combined-ca-bundle\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.702086 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-tls-certs\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.702970 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/970d5293-b621-45e1-a1e4-8cc176c9a148-horizon-secret-key\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.703566 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-secret-key\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.712257 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970d5293-b621-45e1-a1e4-8cc176c9a148-combined-ca-bundle\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.714169 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdsm4\" (UniqueName: \"kubernetes.io/projected/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-kube-api-access-xdsm4\") pod \"horizon-5d466d479d-wdxj6\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.723783 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4wvc\" (UniqueName: \"kubernetes.io/projected/970d5293-b621-45e1-a1e4-8cc176c9a148-kube-api-access-l4wvc\") pod \"horizon-c647fbf5b-7zs9r\" (UID: \"970d5293-b621-45e1-a1e4-8cc176c9a148\") " pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.776290 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:34:54 crc kubenswrapper[4735]: I0215 20:34:54.845164 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:34:55 crc kubenswrapper[4735]: I0215 20:34:55.802954 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:34:55 crc kubenswrapper[4735]: I0215 20:34:55.885543 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-dfdmw"] Feb 15 20:34:55 crc kubenswrapper[4735]: I0215 20:34:55.886476 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="dnsmasq-dns" containerID="cri-o://6d9cdedf8bcecd556655db096a7f647888ea55cf1a0541f7f0672b100320fec3" gracePeriod=10 Feb 15 20:34:56 crc kubenswrapper[4735]: I0215 20:34:56.785338 4735 generic.go:334] "Generic (PLEG): container finished" podID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerID="6d9cdedf8bcecd556655db096a7f647888ea55cf1a0541f7f0672b100320fec3" exitCode=0 Feb 15 20:34:56 crc kubenswrapper[4735]: I0215 20:34:56.785377 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" event={"ID":"fc2f480c-f0b3-4677-ad26-2b4b5071ad44","Type":"ContainerDied","Data":"6d9cdedf8bcecd556655db096a7f647888ea55cf1a0541f7f0672b100320fec3"} Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.555104 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.744937 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-622mp" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.835422 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-622mp" event={"ID":"09c547be-bb78-4d22-adcf-3643dd828a78","Type":"ContainerDied","Data":"5eb9e31a07e40f6abd808d0283e62a2d8822aed876f38c705deb46c560377996"} Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.835461 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eb9e31a07e40f6abd808d0283e62a2d8822aed876f38c705deb46c560377996" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.835508 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-622mp" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.854265 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krpng\" (UniqueName: \"kubernetes.io/projected/09c547be-bb78-4d22-adcf-3643dd828a78-kube-api-access-krpng\") pod \"09c547be-bb78-4d22-adcf-3643dd828a78\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.854315 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-combined-ca-bundle\") pod \"09c547be-bb78-4d22-adcf-3643dd828a78\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.854390 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-scripts\") pod \"09c547be-bb78-4d22-adcf-3643dd828a78\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.856184 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-config-data\") pod \"09c547be-bb78-4d22-adcf-3643dd828a78\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.856229 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-credential-keys\") pod \"09c547be-bb78-4d22-adcf-3643dd828a78\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.856257 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-fernet-keys\") pod \"09c547be-bb78-4d22-adcf-3643dd828a78\" (UID: \"09c547be-bb78-4d22-adcf-3643dd828a78\") " Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.860584 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "09c547be-bb78-4d22-adcf-3643dd828a78" (UID: "09c547be-bb78-4d22-adcf-3643dd828a78"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.862959 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-scripts" (OuterVolumeSpecName: "scripts") pod "09c547be-bb78-4d22-adcf-3643dd828a78" (UID: "09c547be-bb78-4d22-adcf-3643dd828a78"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.870206 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09c547be-bb78-4d22-adcf-3643dd828a78-kube-api-access-krpng" (OuterVolumeSpecName: "kube-api-access-krpng") pod "09c547be-bb78-4d22-adcf-3643dd828a78" (UID: "09c547be-bb78-4d22-adcf-3643dd828a78"). InnerVolumeSpecName "kube-api-access-krpng". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.882273 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "09c547be-bb78-4d22-adcf-3643dd828a78" (UID: "09c547be-bb78-4d22-adcf-3643dd828a78"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.898850 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09c547be-bb78-4d22-adcf-3643dd828a78" (UID: "09c547be-bb78-4d22-adcf-3643dd828a78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.903213 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-config-data" (OuterVolumeSpecName: "config-data") pod "09c547be-bb78-4d22-adcf-3643dd828a78" (UID: "09c547be-bb78-4d22-adcf-3643dd828a78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.960329 4735 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.960356 4735 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.960374 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krpng\" (UniqueName: \"kubernetes.io/projected/09c547be-bb78-4d22-adcf-3643dd828a78-kube-api-access-krpng\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.960387 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.960396 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:00 crc kubenswrapper[4735]: I0215 20:35:00.960404 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09c547be-bb78-4d22-adcf-3643dd828a78-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.873513 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-622mp"] Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.892417 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-622mp"] Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.949299 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4r5s8"] Feb 15 20:35:01 crc kubenswrapper[4735]: E0215 20:35:01.949723 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c547be-bb78-4d22-adcf-3643dd828a78" containerName="keystone-bootstrap" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.949746 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c547be-bb78-4d22-adcf-3643dd828a78" containerName="keystone-bootstrap" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.949939 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="09c547be-bb78-4d22-adcf-3643dd828a78" containerName="keystone-bootstrap" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.950694 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.952942 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.955389 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.955648 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dljv4" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.957239 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.964248 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4r5s8"] Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.969329 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.980383 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-fernet-keys\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.980455 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-scripts\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.980477 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-credential-keys\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.980502 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-config-data\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.980536 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwr89\" (UniqueName: \"kubernetes.io/projected/c924b448-cd94-426c-bc25-ad0efa381777-kube-api-access-dwr89\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:01 crc kubenswrapper[4735]: I0215 20:35:01.980707 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-combined-ca-bundle\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.084221 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-scripts\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.084276 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-credential-keys\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.084308 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-config-data\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.084340 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwr89\" (UniqueName: \"kubernetes.io/projected/c924b448-cd94-426c-bc25-ad0efa381777-kube-api-access-dwr89\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.084433 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-combined-ca-bundle\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.084451 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-fernet-keys\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.089864 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-fernet-keys\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.090519 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-combined-ca-bundle\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.091561 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-config-data\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.094930 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-credential-keys\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.095276 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-scripts\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.105668 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwr89\" (UniqueName: \"kubernetes.io/projected/c924b448-cd94-426c-bc25-ad0efa381777-kube-api-access-dwr89\") pod \"keystone-bootstrap-4r5s8\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.267261 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:02 crc kubenswrapper[4735]: I0215 20:35:02.897116 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09c547be-bb78-4d22-adcf-3643dd828a78" path="/var/lib/kubelet/pods/09c547be-bb78-4d22-adcf-3643dd828a78/volumes" Feb 15 20:35:03 crc kubenswrapper[4735]: E0215 20:35:03.736189 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 15 20:35:03 crc kubenswrapper[4735]: E0215 20:35:03.736341 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kzrk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-ph5qb_openstack(76a8192d-7bcd-4f4a-b798-907d67b3d21c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:35:03 crc kubenswrapper[4735]: E0215 20:35:03.737471 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-ph5qb" podUID="76a8192d-7bcd-4f4a-b798-907d67b3d21c" Feb 15 20:35:03 crc kubenswrapper[4735]: E0215 20:35:03.880753 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-ph5qb" podUID="76a8192d-7bcd-4f4a-b798-907d67b3d21c" Feb 15 20:35:05 crc kubenswrapper[4735]: I0215 20:35:05.554399 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 15 20:35:12 crc kubenswrapper[4735]: I0215 20:35:12.966549 4735 generic.go:334] "Generic (PLEG): container finished" podID="6ae5b3cf-b0c9-4d46-ac26-491df4077db9" containerID="c2bf42173a97f9145aff4eb6ccff1e5f696023d0eaec1c7d82f9d307e916cc36" exitCode=0 Feb 15 20:35:12 crc kubenswrapper[4735]: I0215 20:35:12.966638 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hk7jh" event={"ID":"6ae5b3cf-b0c9-4d46-ac26-491df4077db9","Type":"ContainerDied","Data":"c2bf42173a97f9145aff4eb6ccff1e5f696023d0eaec1c7d82f9d307e916cc36"} Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.555023 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.555411 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.663194 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.669315 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797007 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-config-data\") pod \"d033877b-3f4d-4d60-85da-6b029ad9856c\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797059 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-config-data\") pod \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797108 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-combined-ca-bundle\") pod \"d033877b-3f4d-4d60-85da-6b029ad9856c\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797150 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmbfh\" (UniqueName: \"kubernetes.io/projected/a0f5ac85-f76f-4aee-a16b-df6e2e549276-kube-api-access-tmbfh\") pod \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797197 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"d033877b-3f4d-4d60-85da-6b029ad9856c\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797214 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797248 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-combined-ca-bundle\") pod \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797281 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-logs\") pod \"d033877b-3f4d-4d60-85da-6b029ad9856c\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797307 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thxp6\" (UniqueName: \"kubernetes.io/projected/d033877b-3f4d-4d60-85da-6b029ad9856c-kube-api-access-thxp6\") pod \"d033877b-3f4d-4d60-85da-6b029ad9856c\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797329 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-logs\") pod \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797361 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-public-tls-certs\") pod \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797382 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-scripts\") pod \"d033877b-3f4d-4d60-85da-6b029ad9856c\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797402 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-httpd-run\") pod \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797423 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-scripts\") pod \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\" (UID: \"a0f5ac85-f76f-4aee-a16b-df6e2e549276\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797479 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-internal-tls-certs\") pod \"d033877b-3f4d-4d60-85da-6b029ad9856c\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797498 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-httpd-run\") pod \"d033877b-3f4d-4d60-85da-6b029ad9856c\" (UID: \"d033877b-3f4d-4d60-85da-6b029ad9856c\") " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.797850 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-logs" (OuterVolumeSpecName: "logs") pod "d033877b-3f4d-4d60-85da-6b029ad9856c" (UID: "d033877b-3f4d-4d60-85da-6b029ad9856c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.799063 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a0f5ac85-f76f-4aee-a16b-df6e2e549276" (UID: "a0f5ac85-f76f-4aee-a16b-df6e2e549276"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.799266 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-logs" (OuterVolumeSpecName: "logs") pod "a0f5ac85-f76f-4aee-a16b-df6e2e549276" (UID: "a0f5ac85-f76f-4aee-a16b-df6e2e549276"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.810258 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f5ac85-f76f-4aee-a16b-df6e2e549276-kube-api-access-tmbfh" (OuterVolumeSpecName: "kube-api-access-tmbfh") pod "a0f5ac85-f76f-4aee-a16b-df6e2e549276" (UID: "a0f5ac85-f76f-4aee-a16b-df6e2e549276"). InnerVolumeSpecName "kube-api-access-tmbfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.810432 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "a0f5ac85-f76f-4aee-a16b-df6e2e549276" (UID: "a0f5ac85-f76f-4aee-a16b-df6e2e549276"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.814315 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d033877b-3f4d-4d60-85da-6b029ad9856c" (UID: "d033877b-3f4d-4d60-85da-6b029ad9856c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.816121 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "d033877b-3f4d-4d60-85da-6b029ad9856c" (UID: "d033877b-3f4d-4d60-85da-6b029ad9856c"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.818188 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d033877b-3f4d-4d60-85da-6b029ad9856c-kube-api-access-thxp6" (OuterVolumeSpecName: "kube-api-access-thxp6") pod "d033877b-3f4d-4d60-85da-6b029ad9856c" (UID: "d033877b-3f4d-4d60-85da-6b029ad9856c"). InnerVolumeSpecName "kube-api-access-thxp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.818286 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-scripts" (OuterVolumeSpecName: "scripts") pod "d033877b-3f4d-4d60-85da-6b029ad9856c" (UID: "d033877b-3f4d-4d60-85da-6b029ad9856c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.842053 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-scripts" (OuterVolumeSpecName: "scripts") pod "a0f5ac85-f76f-4aee-a16b-df6e2e549276" (UID: "a0f5ac85-f76f-4aee-a16b-df6e2e549276"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.846338 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0f5ac85-f76f-4aee-a16b-df6e2e549276" (UID: "a0f5ac85-f76f-4aee-a16b-df6e2e549276"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.865405 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d033877b-3f4d-4d60-85da-6b029ad9856c" (UID: "d033877b-3f4d-4d60-85da-6b029ad9856c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.866786 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-config-data" (OuterVolumeSpecName: "config-data") pod "d033877b-3f4d-4d60-85da-6b029ad9856c" (UID: "d033877b-3f4d-4d60-85da-6b029ad9856c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.867646 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a0f5ac85-f76f-4aee-a16b-df6e2e549276" (UID: "a0f5ac85-f76f-4aee-a16b-df6e2e549276"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.869122 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d033877b-3f4d-4d60-85da-6b029ad9856c" (UID: "d033877b-3f4d-4d60-85da-6b029ad9856c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.879767 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-config-data" (OuterVolumeSpecName: "config-data") pod "a0f5ac85-f76f-4aee-a16b-df6e2e549276" (UID: "a0f5ac85-f76f-4aee-a16b-df6e2e549276"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899051 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899085 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmbfh\" (UniqueName: \"kubernetes.io/projected/a0f5ac85-f76f-4aee-a16b-df6e2e549276-kube-api-access-tmbfh\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899113 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899126 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899136 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899145 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899156 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thxp6\" (UniqueName: \"kubernetes.io/projected/d033877b-3f4d-4d60-85da-6b029ad9856c-kube-api-access-thxp6\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899164 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899173 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899181 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899190 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f5ac85-f76f-4aee-a16b-df6e2e549276-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899197 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899205 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899213 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d033877b-3f4d-4d60-85da-6b029ad9856c-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899223 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d033877b-3f4d-4d60-85da-6b029ad9856c-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.899232 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f5ac85-f76f-4aee-a16b-df6e2e549276-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.915136 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.917674 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.991491 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d033877b-3f4d-4d60-85da-6b029ad9856c","Type":"ContainerDied","Data":"e638b7b6e25f2c83739fd6c7b3abe0b1cb91e3cc4aa18a26bec277c1203b4e5a"} Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.991524 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.991537 4735 scope.go:117] "RemoveContainer" containerID="5da788e5749d2834c908ea6d4621ec0d73cec4b3f1e87d736d597e3647d6f8a4" Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.995093 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0f5ac85-f76f-4aee-a16b-df6e2e549276","Type":"ContainerDied","Data":"a4066efa518fb6129f7562c35856c6cd73911e46073d573429419f55d76857be"} Feb 15 20:35:15 crc kubenswrapper[4735]: I0215 20:35:15.995175 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.003600 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.003628 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.044232 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.057919 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.086470 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:35:16 crc kubenswrapper[4735]: E0215 20:35:16.086876 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerName="glance-log" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.086896 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerName="glance-log" Feb 15 20:35:16 crc kubenswrapper[4735]: E0215 20:35:16.086917 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerName="glance-httpd" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.086922 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerName="glance-httpd" Feb 15 20:35:16 crc kubenswrapper[4735]: E0215 20:35:16.086961 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerName="glance-httpd" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.086967 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerName="glance-httpd" Feb 15 20:35:16 crc kubenswrapper[4735]: E0215 20:35:16.086976 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerName="glance-log" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.086983 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerName="glance-log" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.087148 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerName="glance-httpd" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.087156 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerName="glance-log" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.087165 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d033877b-3f4d-4d60-85da-6b029ad9856c" containerName="glance-httpd" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.087181 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" containerName="glance-log" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.088011 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.090144 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.090367 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kfjgg" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.093132 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.093384 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.096509 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.106826 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.136023 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.167172 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.170211 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.172716 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.172875 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.183820 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.206591 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.206639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.206686 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.206770 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.206814 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpj27\" (UniqueName: \"kubernetes.io/projected/1bdadc81-14d1-469b-82c1-68f79cfa71a9-kube-api-access-qpj27\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.206841 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.206865 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.206907 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: E0215 20:35:16.250488 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Feb 15 20:35:16 crc kubenswrapper[4735]: E0215 20:35:16.250636 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d7h558h59fh5fdhf8h55fh5f7h5fbhcch674h595h76h58ch7h579h596h589h5fch687hd7h89h65ch5fch5b9h698h74h8bh96hd7h68h695h5f4q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-62lg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(5d7a00e9-2271-49b1-a74f-51f5f668995a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.256261 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.263886 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308107 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308156 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308196 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308217 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308273 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308326 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-config-data\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308343 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-logs\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308363 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308397 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpj27\" (UniqueName: \"kubernetes.io/projected/1bdadc81-14d1-469b-82c1-68f79cfa71a9-kube-api-access-qpj27\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308421 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308451 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308473 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-scripts\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308494 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308516 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk4zs\" (UniqueName: \"kubernetes.io/projected/e48a6f8b-a91e-438b-832b-b9ba15c87e15-kube-api-access-mk4zs\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308542 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.308584 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.309656 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.310376 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.321555 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.324104 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.327320 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.330311 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpj27\" (UniqueName: \"kubernetes.io/projected/1bdadc81-14d1-469b-82c1-68f79cfa71a9-kube-api-access-qpj27\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.331489 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.332341 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.372160 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.406550 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.409920 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-config\") pod \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410048 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-config\") pod \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410220 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plkhg\" (UniqueName: \"kubernetes.io/projected/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-kube-api-access-plkhg\") pod \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410265 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-sb\") pod \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410298 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-nb\") pod \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410365 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-swift-storage-0\") pod \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410391 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-combined-ca-bundle\") pod \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\" (UID: \"6ae5b3cf-b0c9-4d46-ac26-491df4077db9\") " Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410468 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b852\" (UniqueName: \"kubernetes.io/projected/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-kube-api-access-6b852\") pod \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410514 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-svc\") pod \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\" (UID: \"fc2f480c-f0b3-4677-ad26-2b4b5071ad44\") " Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410890 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-config-data\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410910 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-logs\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.410983 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.411023 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-scripts\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.411056 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk4zs\" (UniqueName: \"kubernetes.io/projected/e48a6f8b-a91e-438b-832b-b9ba15c87e15-kube-api-access-mk4zs\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.411087 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.411187 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.411507 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-logs\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.411664 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.412183 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.416781 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-kube-api-access-plkhg" (OuterVolumeSpecName: "kube-api-access-plkhg") pod "6ae5b3cf-b0c9-4d46-ac26-491df4077db9" (UID: "6ae5b3cf-b0c9-4d46-ac26-491df4077db9"). InnerVolumeSpecName "kube-api-access-plkhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.427935 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-kube-api-access-6b852" (OuterVolumeSpecName: "kube-api-access-6b852") pod "fc2f480c-f0b3-4677-ad26-2b4b5071ad44" (UID: "fc2f480c-f0b3-4677-ad26-2b4b5071ad44"). InnerVolumeSpecName "kube-api-access-6b852". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.428218 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.428250 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-scripts\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.429084 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-config-data\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.432741 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.439701 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk4zs\" (UniqueName: \"kubernetes.io/projected/e48a6f8b-a91e-438b-832b-b9ba15c87e15-kube-api-access-mk4zs\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.461020 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-config" (OuterVolumeSpecName: "config") pod "6ae5b3cf-b0c9-4d46-ac26-491df4077db9" (UID: "6ae5b3cf-b0c9-4d46-ac26-491df4077db9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.465714 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ae5b3cf-b0c9-4d46-ac26-491df4077db9" (UID: "6ae5b3cf-b0c9-4d46-ac26-491df4077db9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.482620 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.483681 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fc2f480c-f0b3-4677-ad26-2b4b5071ad44" (UID: "fc2f480c-f0b3-4677-ad26-2b4b5071ad44"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.499763 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fc2f480c-f0b3-4677-ad26-2b4b5071ad44" (UID: "fc2f480c-f0b3-4677-ad26-2b4b5071ad44"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.505435 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fc2f480c-f0b3-4677-ad26-2b4b5071ad44" (UID: "fc2f480c-f0b3-4677-ad26-2b4b5071ad44"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.509768 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-config" (OuterVolumeSpecName: "config") pod "fc2f480c-f0b3-4677-ad26-2b4b5071ad44" (UID: "fc2f480c-f0b3-4677-ad26-2b4b5071ad44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.513982 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plkhg\" (UniqueName: \"kubernetes.io/projected/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-kube-api-access-plkhg\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.514011 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.514021 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.514069 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.514078 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.514086 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b852\" (UniqueName: \"kubernetes.io/projected/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-kube-api-access-6b852\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.514095 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ae5b3cf-b0c9-4d46-ac26-491df4077db9-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.514107 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.517116 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fc2f480c-f0b3-4677-ad26-2b4b5071ad44" (UID: "fc2f480c-f0b3-4677-ad26-2b4b5071ad44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.615815 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2f480c-f0b3-4677-ad26-2b4b5071ad44-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.785071 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.909857 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f5ac85-f76f-4aee-a16b-df6e2e549276" path="/var/lib/kubelet/pods/a0f5ac85-f76f-4aee-a16b-df6e2e549276/volumes" Feb 15 20:35:16 crc kubenswrapper[4735]: I0215 20:35:16.911600 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d033877b-3f4d-4d60-85da-6b029ad9856c" path="/var/lib/kubelet/pods/d033877b-3f4d-4d60-85da-6b029ad9856c/volumes" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.038237 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" event={"ID":"fc2f480c-f0b3-4677-ad26-2b4b5071ad44","Type":"ContainerDied","Data":"4838722be565f495ceac73db949cb7d656d47af2da5a40254006801cf26db1d6"} Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.038323 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.050501 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hk7jh" event={"ID":"6ae5b3cf-b0c9-4d46-ac26-491df4077db9","Type":"ContainerDied","Data":"b62a09caa198afd75e115b735a1559c1d01429c5c7a5e616d2e3cef5d47a7a9d"} Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.050536 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b62a09caa198afd75e115b735a1559c1d01429c5c7a5e616d2e3cef5d47a7a9d" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.050590 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hk7jh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.063398 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-dfdmw"] Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.081651 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-dfdmw"] Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.505597 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8ggxh"] Feb 15 20:35:17 crc kubenswrapper[4735]: E0215 20:35:17.506152 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="init" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.506191 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="init" Feb 15 20:35:17 crc kubenswrapper[4735]: E0215 20:35:17.506207 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="dnsmasq-dns" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.506212 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="dnsmasq-dns" Feb 15 20:35:17 crc kubenswrapper[4735]: E0215 20:35:17.506231 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae5b3cf-b0c9-4d46-ac26-491df4077db9" containerName="neutron-db-sync" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.506280 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae5b3cf-b0c9-4d46-ac26-491df4077db9" containerName="neutron-db-sync" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.506497 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="dnsmasq-dns" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.506535 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae5b3cf-b0c9-4d46-ac26-491df4077db9" containerName="neutron-db-sync" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.507609 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.527056 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8ggxh"] Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.637775 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.637857 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-svc\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.637905 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plmt6\" (UniqueName: \"kubernetes.io/projected/ed582df2-f4b9-4c7d-9f67-695b388d649d-kube-api-access-plmt6\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.637936 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.638055 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.638084 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-config\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.647466 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-59f96fd876-q78bj"] Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.648809 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.658697 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.658859 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.659064 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-w7s2g" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.659170 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.665435 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59f96fd876-q78bj"] Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739342 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl48k\" (UniqueName: \"kubernetes.io/projected/4e551f68-1918-4b4a-99de-c7e05c507187-kube-api-access-vl48k\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739385 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-httpd-config\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739416 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739433 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-config\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739476 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-svc\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739502 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plmt6\" (UniqueName: \"kubernetes.io/projected/ed582df2-f4b9-4c7d-9f67-695b388d649d-kube-api-access-plmt6\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739530 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-combined-ca-bundle\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739548 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-ovndb-tls-certs\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739568 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739593 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.739618 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-config\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.741223 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-config\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.741901 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-svc\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.743425 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.743851 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.744356 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.775242 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plmt6\" (UniqueName: \"kubernetes.io/projected/ed582df2-f4b9-4c7d-9f67-695b388d649d-kube-api-access-plmt6\") pod \"dnsmasq-dns-6b7b667979-8ggxh\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.828845 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.841870 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-combined-ca-bundle\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.841926 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-ovndb-tls-certs\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.842115 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl48k\" (UniqueName: \"kubernetes.io/projected/4e551f68-1918-4b4a-99de-c7e05c507187-kube-api-access-vl48k\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.842940 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-httpd-config\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.843021 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-config\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.847091 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-config\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.849458 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-ovndb-tls-certs\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.856372 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-combined-ca-bundle\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.866286 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-httpd-config\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.885649 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl48k\" (UniqueName: \"kubernetes.io/projected/4e551f68-1918-4b4a-99de-c7e05c507187-kube-api-access-vl48k\") pod \"neutron-59f96fd876-q78bj\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:17 crc kubenswrapper[4735]: I0215 20:35:17.988426 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:18 crc kubenswrapper[4735]: E0215 20:35:18.739396 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 15 20:35:18 crc kubenswrapper[4735]: E0215 20:35:18.739828 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kxvj5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-6htdh_openstack(ec4fe5e3-ca92-4814-8921-62117a037281): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 20:35:18 crc kubenswrapper[4735]: E0215 20:35:18.741234 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-6htdh" podUID="ec4fe5e3-ca92-4814-8921-62117a037281" Feb 15 20:35:18 crc kubenswrapper[4735]: I0215 20:35:18.768828 4735 scope.go:117] "RemoveContainer" containerID="14f6a3590b5e07b06907bf5f440089060cab01534838fd703d5bf1f926bea989" Feb 15 20:35:18 crc kubenswrapper[4735]: I0215 20:35:18.945506 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" path="/var/lib/kubelet/pods/fc2f480c-f0b3-4677-ad26-2b4b5071ad44/volumes" Feb 15 20:35:18 crc kubenswrapper[4735]: I0215 20:35:18.988194 4735 scope.go:117] "RemoveContainer" containerID="98a6d1826e7109483ece50a3320347909ca3a08afe1fb8c7248a2b48ac3087cf" Feb 15 20:35:19 crc kubenswrapper[4735]: E0215 20:35:19.162460 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-6htdh" podUID="ec4fe5e3-ca92-4814-8921-62117a037281" Feb 15 20:35:19 crc kubenswrapper[4735]: I0215 20:35:19.236653 4735 scope.go:117] "RemoveContainer" containerID="e0ddeb09f9aca4f4a828cb390e5fc2c69d35b9fb9ffefe19b41667129797535c" Feb 15 20:35:19 crc kubenswrapper[4735]: I0215 20:35:19.346170 4735 scope.go:117] "RemoveContainer" containerID="6d9cdedf8bcecd556655db096a7f647888ea55cf1a0541f7f0672b100320fec3" Feb 15 20:35:19 crc kubenswrapper[4735]: I0215 20:35:19.397259 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c647fbf5b-7zs9r"] Feb 15 20:35:19 crc kubenswrapper[4735]: I0215 20:35:19.420959 4735 scope.go:117] "RemoveContainer" containerID="b6ac5646db598059deeb27c0b0edbf68ee2c5051dc54d264c2fdff563aceecb9" Feb 15 20:35:19 crc kubenswrapper[4735]: I0215 20:35:19.493020 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d466d479d-wdxj6"] Feb 15 20:35:19 crc kubenswrapper[4735]: I0215 20:35:19.679653 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:35:19 crc kubenswrapper[4735]: I0215 20:35:19.679716 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.029257 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4r5s8"] Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.096750 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-557bd9599f-5lqht"] Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.105812 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.110969 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.111114 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.148245 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ph5qb" event={"ID":"76a8192d-7bcd-4f4a-b798-907d67b3d21c","Type":"ContainerStarted","Data":"410a4036f6910ec4b0f837c8cfff0e0d7d3938ae03838de515d272f46e927a52"} Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.169034 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8ggxh"] Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.180455 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-557bd9599f-5lqht"] Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.207092 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-combined-ca-bundle\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.207191 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdksf\" (UniqueName: \"kubernetes.io/projected/2ef1315b-689a-4bab-9e16-0f767cd22170-kube-api-access-vdksf\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.207295 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-config\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.207344 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-internal-tls-certs\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.207374 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-httpd-config\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.207410 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-public-tls-certs\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.207428 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-ovndb-tls-certs\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.242831 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pss7t" event={"ID":"327449fc-a14d-4594-b8a8-c758aeebec34","Type":"ContainerStarted","Data":"09f9578052a97aa9ebe8cfb14289319207556a64165f83e900c09158d6b673d2"} Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.295016 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74dd7854c7-t2t6r" event={"ID":"faf23e04-da47-4121-9e55-d003d8d6dcfe","Type":"ContainerStarted","Data":"f295535d4a74689debfd09b839f03d0ba86e654fdda24a7807dcd72de4f9eb23"} Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.314927 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-public-tls-certs\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.314981 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-ovndb-tls-certs\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.315009 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-combined-ca-bundle\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.316686 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdksf\" (UniqueName: \"kubernetes.io/projected/2ef1315b-689a-4bab-9e16-0f767cd22170-kube-api-access-vdksf\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.317493 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-config\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.317543 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-internal-tls-certs\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.317567 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-httpd-config\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.333813 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6898675dc5-fh44t" event={"ID":"a6f89bfa-6fb3-40d0-9741-cfddddb9a685","Type":"ContainerStarted","Data":"5c3c8f494e812ff4a8c0aee1120fe0bd32c14ec07a9effb34a54ef3370575760"} Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.352832 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-httpd-config\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.371890 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ph5qb" podStartSLOduration=3.695209169 podStartE2EDuration="35.371868356s" podCreationTimestamp="2026-02-15 20:34:45 +0000 UTC" firstStartedPulling="2026-02-15 20:34:47.560021844 +0000 UTC m=+1095.426037467" lastFinishedPulling="2026-02-15 20:35:19.236681031 +0000 UTC m=+1127.102696654" observedRunningTime="2026-02-15 20:35:20.295479915 +0000 UTC m=+1128.161495538" watchObservedRunningTime="2026-02-15 20:35:20.371868356 +0000 UTC m=+1128.237883979" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.412002 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerStarted","Data":"b1fff493b8f3f991759d71a9f7417a87ca0a4d350b5f0355919487363eeb3c7e"} Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.421116 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.424110 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4r5s8" event={"ID":"c924b448-cd94-426c-bc25-ad0efa381777","Type":"ContainerStarted","Data":"e4a0681ad006bfd509754a0f839b982c75ada18843c18536b9b0ddd4eb41fa57"} Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.435116 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c647fbf5b-7zs9r" event={"ID":"970d5293-b621-45e1-a1e4-8cc176c9a148","Type":"ContainerStarted","Data":"cf924bc92703446b1cb7f02e4e2fa336a5afac7af57e96a35dca32867ac38145"} Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.450535 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-public-tls-certs\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.452008 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-pss7t" podStartSLOduration=6.718614573 podStartE2EDuration="35.451989177s" podCreationTimestamp="2026-02-15 20:34:45 +0000 UTC" firstStartedPulling="2026-02-15 20:34:46.821632263 +0000 UTC m=+1094.687647886" lastFinishedPulling="2026-02-15 20:35:15.555006867 +0000 UTC m=+1123.421022490" observedRunningTime="2026-02-15 20:35:20.331103606 +0000 UTC m=+1128.197119229" watchObservedRunningTime="2026-02-15 20:35:20.451989177 +0000 UTC m=+1128.318004800" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.452479 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f7d59545-hbprm" event={"ID":"1be12817-7c08-4d66-938d-98ce9f79e4fd","Type":"ContainerStarted","Data":"10a22afe76edb65bfb324a00acf60c68c9864393f0fe03999f09757222e33183"} Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.452515 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f7d59545-hbprm" event={"ID":"1be12817-7c08-4d66-938d-98ce9f79e4fd","Type":"ContainerStarted","Data":"81739c66d31bfab9b2a5e97080b0eccbabd62978dd60f71980b7bfc6c8bc8839"} Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.452625 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74f7d59545-hbprm" podUID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerName="horizon-log" containerID="cri-o://81739c66d31bfab9b2a5e97080b0eccbabd62978dd60f71980b7bfc6c8bc8839" gracePeriod=30 Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.453852 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74f7d59545-hbprm" podUID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerName="horizon" containerID="cri-o://10a22afe76edb65bfb324a00acf60c68c9864393f0fe03999f09757222e33183" gracePeriod=30 Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.463832 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-internal-tls-certs\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.464426 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-config\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.476022 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-combined-ca-bundle\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.484766 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-ovndb-tls-certs\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.495702 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdksf\" (UniqueName: \"kubernetes.io/projected/2ef1315b-689a-4bab-9e16-0f767cd22170-kube-api-access-vdksf\") pod \"neutron-557bd9599f-5lqht\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.514240 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.525541 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74f7d59545-hbprm" podStartSLOduration=5.361848063 podStartE2EDuration="36.525522771s" podCreationTimestamp="2026-02-15 20:34:44 +0000 UTC" firstStartedPulling="2026-02-15 20:34:47.526259783 +0000 UTC m=+1095.392275406" lastFinishedPulling="2026-02-15 20:35:18.689934501 +0000 UTC m=+1126.555950114" observedRunningTime="2026-02-15 20:35:20.509797887 +0000 UTC m=+1128.375813510" watchObservedRunningTime="2026-02-15 20:35:20.525522771 +0000 UTC m=+1128.391538394" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.557059 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-dfdmw" podUID="fc2f480c-f0b3-4677-ad26-2b4b5071ad44" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.628565 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59f96fd876-q78bj"] Feb 15 20:35:20 crc kubenswrapper[4735]: I0215 20:35:20.653477 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:21 crc kubenswrapper[4735]: W0215 20:35:21.289926 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e551f68_1918_4b4a_99de_c7e05c507187.slice/crio-ce1d49f8915c985e900f99b17ba2c54639fa8577c4f44b083ef75d73f58294d4 WatchSource:0}: Error finding container ce1d49f8915c985e900f99b17ba2c54639fa8577c4f44b083ef75d73f58294d4: Status 404 returned error can't find the container with id ce1d49f8915c985e900f99b17ba2c54639fa8577c4f44b083ef75d73f58294d4 Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.479716 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4r5s8" event={"ID":"c924b448-cd94-426c-bc25-ad0efa381777","Type":"ContainerStarted","Data":"8fdb6b04f4a90b71fd7b845d209a2074d7f9c23207a44a099f78922ae5354d53"} Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.491348 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c647fbf5b-7zs9r" event={"ID":"970d5293-b621-45e1-a1e4-8cc176c9a148","Type":"ContainerStarted","Data":"0bcc8bcaa40eaec65c475349232babc487528aeb72575254d5a9333fa0ae1da3"} Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.508429 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" event={"ID":"ed582df2-f4b9-4c7d-9f67-695b388d649d","Type":"ContainerStarted","Data":"95deee861412ac7aaf695eda38712435721b54c678d11f79608b7f2f85e4ffb2"} Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.511041 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bdadc81-14d1-469b-82c1-68f79cfa71a9","Type":"ContainerStarted","Data":"55361a0c531f4e64c8e96845063e9a885a0d31dc1a08b49bbe8b89f8c874ea89"} Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.515935 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4r5s8" podStartSLOduration=20.51591769 podStartE2EDuration="20.51591769s" podCreationTimestamp="2026-02-15 20:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:21.504698947 +0000 UTC m=+1129.370714570" watchObservedRunningTime="2026-02-15 20:35:21.51591769 +0000 UTC m=+1129.381933313" Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.528656 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerStarted","Data":"5d80cb6d2dc2ccf6ac05974c0b49e10cfa35444bc235b804bde522b9f6bb18fb"} Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.532895 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59f96fd876-q78bj" event={"ID":"4e551f68-1918-4b4a-99de-c7e05c507187","Type":"ContainerStarted","Data":"ce1d49f8915c985e900f99b17ba2c54639fa8577c4f44b083ef75d73f58294d4"} Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.537247 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e48a6f8b-a91e-438b-832b-b9ba15c87e15","Type":"ContainerStarted","Data":"cad371487c069f9da3937fb49038d5395b203f526bda6393332258698c6ee339"} Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.538613 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74dd7854c7-t2t6r" event={"ID":"faf23e04-da47-4121-9e55-d003d8d6dcfe","Type":"ContainerStarted","Data":"363de19826e86b82ff64c234ca3ec5e5b9bfa1a8265f4b25fb7531f400e00c7c"} Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.538749 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74dd7854c7-t2t6r" podUID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerName="horizon-log" containerID="cri-o://f295535d4a74689debfd09b839f03d0ba86e654fdda24a7807dcd72de4f9eb23" gracePeriod=30 Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.539234 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74dd7854c7-t2t6r" podUID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerName="horizon" containerID="cri-o://363de19826e86b82ff64c234ca3ec5e5b9bfa1a8265f4b25fb7531f400e00c7c" gracePeriod=30 Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.570281 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74dd7854c7-t2t6r" podStartSLOduration=5.269020627 podStartE2EDuration="36.570261476s" podCreationTimestamp="2026-02-15 20:34:45 +0000 UTC" firstStartedPulling="2026-02-15 20:34:47.598678387 +0000 UTC m=+1095.464694020" lastFinishedPulling="2026-02-15 20:35:18.899919246 +0000 UTC m=+1126.765934869" observedRunningTime="2026-02-15 20:35:21.558428466 +0000 UTC m=+1129.424444089" watchObservedRunningTime="2026-02-15 20:35:21.570261476 +0000 UTC m=+1129.436277099" Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.581814 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6898675dc5-fh44t" podUID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerName="horizon-log" containerID="cri-o://5c3c8f494e812ff4a8c0aee1120fe0bd32c14ec07a9effb34a54ef3370575760" gracePeriod=30 Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.581866 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6898675dc5-fh44t" event={"ID":"a6f89bfa-6fb3-40d0-9741-cfddddb9a685","Type":"ContainerStarted","Data":"a3e86739b52ad23dbac1a7f4cfcf2a070a9b0f25ee1f6dad3b0794fcaa858d61"} Feb 15 20:35:21 crc kubenswrapper[4735]: I0215 20:35:21.582523 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6898675dc5-fh44t" podUID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerName="horizon" containerID="cri-o://a3e86739b52ad23dbac1a7f4cfcf2a070a9b0f25ee1f6dad3b0794fcaa858d61" gracePeriod=30 Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.077067 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6898675dc5-fh44t" podStartSLOduration=5.431208956 podStartE2EDuration="35.077047647s" podCreationTimestamp="2026-02-15 20:34:47 +0000 UTC" firstStartedPulling="2026-02-15 20:34:49.25573849 +0000 UTC m=+1097.121754113" lastFinishedPulling="2026-02-15 20:35:18.901577191 +0000 UTC m=+1126.767592804" observedRunningTime="2026-02-15 20:35:21.627160461 +0000 UTC m=+1129.493176104" watchObservedRunningTime="2026-02-15 20:35:22.077047647 +0000 UTC m=+1129.943063270" Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.092290 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-557bd9599f-5lqht"] Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.596128 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557bd9599f-5lqht" event={"ID":"2ef1315b-689a-4bab-9e16-0f767cd22170","Type":"ContainerStarted","Data":"7a7b24352b641cb77fad01a59540c8459a7c1234dcf8766c7319d7e46daa5006"} Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.596535 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557bd9599f-5lqht" event={"ID":"2ef1315b-689a-4bab-9e16-0f767cd22170","Type":"ContainerStarted","Data":"5d8c20b9735ab059862611d7ddf724fa4ba532cbd97efd1200d69b4ff9693179"} Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.607356 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerStarted","Data":"509c549dfe02bd37ca623752f3760afb9e5b4c41da36b238f99fb78d66aac256"} Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.619718 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c647fbf5b-7zs9r" event={"ID":"970d5293-b621-45e1-a1e4-8cc176c9a148","Type":"ContainerStarted","Data":"063addf10589b9afd3c588e4171de6d504c19a3cbf44f266f86a25dfe5ca99e3"} Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.625683 4735 generic.go:334] "Generic (PLEG): container finished" podID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerID="82c48854bc164c771936514231eff3628e6b2bc61416220bc8cee0f4bab04980" exitCode=0 Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.625742 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" event={"ID":"ed582df2-f4b9-4c7d-9f67-695b388d649d","Type":"ContainerDied","Data":"82c48854bc164c771936514231eff3628e6b2bc61416220bc8cee0f4bab04980"} Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.636663 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5d466d479d-wdxj6" podStartSLOduration=28.636648604 podStartE2EDuration="28.636648604s" podCreationTimestamp="2026-02-15 20:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:22.629502132 +0000 UTC m=+1130.495517765" watchObservedRunningTime="2026-02-15 20:35:22.636648604 +0000 UTC m=+1130.502664227" Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.638518 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59f96fd876-q78bj" event={"ID":"4e551f68-1918-4b4a-99de-c7e05c507187","Type":"ContainerStarted","Data":"51c478a3412d9dd59ac3624eb9e78d9ded5143fcdf90a0f76ed2fc9271a76bf1"} Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.638551 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59f96fd876-q78bj" event={"ID":"4e551f68-1918-4b4a-99de-c7e05c507187","Type":"ContainerStarted","Data":"7c06fc1659d45c78cd1d29f4b44200f8c4675dc45011376308cfbc0b5abfbba7"} Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.639249 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.642174 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d7a00e9-2271-49b1-a74f-51f5f668995a","Type":"ContainerStarted","Data":"f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99"} Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.709953 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c647fbf5b-7zs9r" podStartSLOduration=28.709916181 podStartE2EDuration="28.709916181s" podCreationTimestamp="2026-02-15 20:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:22.688010971 +0000 UTC m=+1130.554026614" watchObservedRunningTime="2026-02-15 20:35:22.709916181 +0000 UTC m=+1130.575931804" Feb 15 20:35:22 crc kubenswrapper[4735]: I0215 20:35:22.760975 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-59f96fd876-q78bj" podStartSLOduration=5.760956818 podStartE2EDuration="5.760956818s" podCreationTimestamp="2026-02-15 20:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:22.720850256 +0000 UTC m=+1130.586865879" watchObservedRunningTime="2026-02-15 20:35:22.760956818 +0000 UTC m=+1130.626972441" Feb 15 20:35:23 crc kubenswrapper[4735]: I0215 20:35:23.660935 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bdadc81-14d1-469b-82c1-68f79cfa71a9","Type":"ContainerStarted","Data":"0eceef46b722125a7a67a2e0b988daa2f542ecc7fd172b767ea9348df61f5f01"} Feb 15 20:35:23 crc kubenswrapper[4735]: I0215 20:35:23.675729 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557bd9599f-5lqht" event={"ID":"2ef1315b-689a-4bab-9e16-0f767cd22170","Type":"ContainerStarted","Data":"eb9e7b1d420557821175350c51326aa947abdf4c0c45729785ed5d87a383f30c"} Feb 15 20:35:23 crc kubenswrapper[4735]: I0215 20:35:23.676742 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:23 crc kubenswrapper[4735]: I0215 20:35:23.695399 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" event={"ID":"ed582df2-f4b9-4c7d-9f67-695b388d649d","Type":"ContainerStarted","Data":"dfe66441eaccfa6b08bdac1ca3e8b4ad52dc69c9543ac02fe7622c5f3c0ce6b0"} Feb 15 20:35:23 crc kubenswrapper[4735]: I0215 20:35:23.695975 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:23 crc kubenswrapper[4735]: I0215 20:35:23.699916 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e48a6f8b-a91e-438b-832b-b9ba15c87e15","Type":"ContainerStarted","Data":"7f220f4630db2ac2af4c43eee7e69a114fc9d4b048e5ff25b90aaa27e470205d"} Feb 15 20:35:23 crc kubenswrapper[4735]: I0215 20:35:23.724800 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-557bd9599f-5lqht" podStartSLOduration=4.72478484 podStartE2EDuration="4.72478484s" podCreationTimestamp="2026-02-15 20:35:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:23.718675375 +0000 UTC m=+1131.584690998" watchObservedRunningTime="2026-02-15 20:35:23.72478484 +0000 UTC m=+1131.590800463" Feb 15 20:35:23 crc kubenswrapper[4735]: I0215 20:35:23.797416 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" podStartSLOduration=6.79739909 podStartE2EDuration="6.79739909s" podCreationTimestamp="2026-02-15 20:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:23.772131308 +0000 UTC m=+1131.638146931" watchObservedRunningTime="2026-02-15 20:35:23.79739909 +0000 UTC m=+1131.663414713" Feb 15 20:35:24 crc kubenswrapper[4735]: I0215 20:35:24.710507 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e48a6f8b-a91e-438b-832b-b9ba15c87e15","Type":"ContainerStarted","Data":"fce032895e4ab9a42d9c005265ba71abf97dc38fd3a315afabcc1cd553214a3e"} Feb 15 20:35:24 crc kubenswrapper[4735]: I0215 20:35:24.716480 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bdadc81-14d1-469b-82c1-68f79cfa71a9","Type":"ContainerStarted","Data":"2362903ea75605e58912d9ab9e9910e689415b6e37485ede12f5837307f8d68a"} Feb 15 20:35:24 crc kubenswrapper[4735]: I0215 20:35:24.769195 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.769180425 podStartE2EDuration="8.769180425s" podCreationTimestamp="2026-02-15 20:35:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:24.739204907 +0000 UTC m=+1132.605220540" watchObservedRunningTime="2026-02-15 20:35:24.769180425 +0000 UTC m=+1132.635196048" Feb 15 20:35:24 crc kubenswrapper[4735]: I0215 20:35:24.770180 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.770174683 podStartE2EDuration="8.770174683s" podCreationTimestamp="2026-02-15 20:35:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:24.76526475 +0000 UTC m=+1132.631280373" watchObservedRunningTime="2026-02-15 20:35:24.770174683 +0000 UTC m=+1132.636190306" Feb 15 20:35:24 crc kubenswrapper[4735]: I0215 20:35:24.777998 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:35:24 crc kubenswrapper[4735]: I0215 20:35:24.778076 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:35:24 crc kubenswrapper[4735]: I0215 20:35:24.846814 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:35:24 crc kubenswrapper[4735]: I0215 20:35:24.846867 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:35:25 crc kubenswrapper[4735]: I0215 20:35:25.432168 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:35:25 crc kubenswrapper[4735]: I0215 20:35:25.699002 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.407000 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.407043 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.477527 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.495483 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.751176 4735 generic.go:334] "Generic (PLEG): container finished" podID="327449fc-a14d-4594-b8a8-c758aeebec34" containerID="09f9578052a97aa9ebe8cfb14289319207556a64165f83e900c09158d6b673d2" exitCode=0 Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.752261 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pss7t" event={"ID":"327449fc-a14d-4594-b8a8-c758aeebec34","Type":"ContainerDied","Data":"09f9578052a97aa9ebe8cfb14289319207556a64165f83e900c09158d6b673d2"} Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.752580 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.752596 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.786283 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.786334 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.846689 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 15 20:35:26 crc kubenswrapper[4735]: I0215 20:35:26.850464 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 15 20:35:27 crc kubenswrapper[4735]: I0215 20:35:27.761725 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 15 20:35:27 crc kubenswrapper[4735]: I0215 20:35:27.761766 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 15 20:35:28 crc kubenswrapper[4735]: I0215 20:35:28.508833 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:35:29 crc kubenswrapper[4735]: I0215 20:35:29.777070 4735 generic.go:334] "Generic (PLEG): container finished" podID="c924b448-cd94-426c-bc25-ad0efa381777" containerID="8fdb6b04f4a90b71fd7b845d209a2074d7f9c23207a44a099f78922ae5354d53" exitCode=0 Feb 15 20:35:29 crc kubenswrapper[4735]: I0215 20:35:29.778246 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4r5s8" event={"ID":"c924b448-cd94-426c-bc25-ad0efa381777","Type":"ContainerDied","Data":"8fdb6b04f4a90b71fd7b845d209a2074d7f9c23207a44a099f78922ae5354d53"} Feb 15 20:35:29 crc kubenswrapper[4735]: I0215 20:35:29.780822 4735 generic.go:334] "Generic (PLEG): container finished" podID="76a8192d-7bcd-4f4a-b798-907d67b3d21c" containerID="410a4036f6910ec4b0f837c8cfff0e0d7d3938ae03838de515d272f46e927a52" exitCode=0 Feb 15 20:35:29 crc kubenswrapper[4735]: I0215 20:35:29.780902 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ph5qb" event={"ID":"76a8192d-7bcd-4f4a-b798-907d67b3d21c","Type":"ContainerDied","Data":"410a4036f6910ec4b0f837c8cfff0e0d7d3938ae03838de515d272f46e927a52"} Feb 15 20:35:32 crc kubenswrapper[4735]: I0215 20:35:32.319378 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:32 crc kubenswrapper[4735]: I0215 20:35:32.319986 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:35:32 crc kubenswrapper[4735]: I0215 20:35:32.326383 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 15 20:35:32 crc kubenswrapper[4735]: I0215 20:35:32.377380 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 15 20:35:32 crc kubenswrapper[4735]: I0215 20:35:32.831100 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.033256 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-s945p"] Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.033479 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" podUID="da52e1cb-1436-4b18-a2b6-5032d1018b0b" containerName="dnsmasq-dns" containerID="cri-o://046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12" gracePeriod=10 Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.389191 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.405719 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.453496 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pss7t" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610078 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-config-data\") pod \"327449fc-a14d-4594-b8a8-c758aeebec34\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610409 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-combined-ca-bundle\") pod \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610436 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzrk6\" (UniqueName: \"kubernetes.io/projected/76a8192d-7bcd-4f4a-b798-907d67b3d21c-kube-api-access-kzrk6\") pod \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610489 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327449fc-a14d-4594-b8a8-c758aeebec34-logs\") pod \"327449fc-a14d-4594-b8a8-c758aeebec34\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610528 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-fernet-keys\") pod \"c924b448-cd94-426c-bc25-ad0efa381777\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610547 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvbgv\" (UniqueName: \"kubernetes.io/projected/327449fc-a14d-4594-b8a8-c758aeebec34-kube-api-access-zvbgv\") pod \"327449fc-a14d-4594-b8a8-c758aeebec34\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610566 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwr89\" (UniqueName: \"kubernetes.io/projected/c924b448-cd94-426c-bc25-ad0efa381777-kube-api-access-dwr89\") pod \"c924b448-cd94-426c-bc25-ad0efa381777\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610617 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-scripts\") pod \"c924b448-cd94-426c-bc25-ad0efa381777\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610670 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-scripts\") pod \"327449fc-a14d-4594-b8a8-c758aeebec34\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610696 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-combined-ca-bundle\") pod \"c924b448-cd94-426c-bc25-ad0efa381777\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610717 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-credential-keys\") pod \"c924b448-cd94-426c-bc25-ad0efa381777\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610738 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-combined-ca-bundle\") pod \"327449fc-a14d-4594-b8a8-c758aeebec34\" (UID: \"327449fc-a14d-4594-b8a8-c758aeebec34\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610792 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-db-sync-config-data\") pod \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\" (UID: \"76a8192d-7bcd-4f4a-b798-907d67b3d21c\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.610809 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-config-data\") pod \"c924b448-cd94-426c-bc25-ad0efa381777\" (UID: \"c924b448-cd94-426c-bc25-ad0efa381777\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.616094 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/327449fc-a14d-4594-b8a8-c758aeebec34-logs" (OuterVolumeSpecName: "logs") pod "327449fc-a14d-4594-b8a8-c758aeebec34" (UID: "327449fc-a14d-4594-b8a8-c758aeebec34"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.632741 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c924b448-cd94-426c-bc25-ad0efa381777-kube-api-access-dwr89" (OuterVolumeSpecName: "kube-api-access-dwr89") pod "c924b448-cd94-426c-bc25-ad0efa381777" (UID: "c924b448-cd94-426c-bc25-ad0efa381777"). InnerVolumeSpecName "kube-api-access-dwr89". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.634224 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/327449fc-a14d-4594-b8a8-c758aeebec34-kube-api-access-zvbgv" (OuterVolumeSpecName: "kube-api-access-zvbgv") pod "327449fc-a14d-4594-b8a8-c758aeebec34" (UID: "327449fc-a14d-4594-b8a8-c758aeebec34"). InnerVolumeSpecName "kube-api-access-zvbgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.636067 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-scripts" (OuterVolumeSpecName: "scripts") pod "c924b448-cd94-426c-bc25-ad0efa381777" (UID: "c924b448-cd94-426c-bc25-ad0efa381777"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.636224 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a8192d-7bcd-4f4a-b798-907d67b3d21c-kube-api-access-kzrk6" (OuterVolumeSpecName: "kube-api-access-kzrk6") pod "76a8192d-7bcd-4f4a-b798-907d67b3d21c" (UID: "76a8192d-7bcd-4f4a-b798-907d67b3d21c"). InnerVolumeSpecName "kube-api-access-kzrk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.645073 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c924b448-cd94-426c-bc25-ad0efa381777" (UID: "c924b448-cd94-426c-bc25-ad0efa381777"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.646282 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c924b448-cd94-426c-bc25-ad0efa381777" (UID: "c924b448-cd94-426c-bc25-ad0efa381777"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.650108 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-scripts" (OuterVolumeSpecName: "scripts") pod "327449fc-a14d-4594-b8a8-c758aeebec34" (UID: "327449fc-a14d-4594-b8a8-c758aeebec34"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.666511 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "76a8192d-7bcd-4f4a-b798-907d67b3d21c" (UID: "76a8192d-7bcd-4f4a-b798-907d67b3d21c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.707742 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.713801 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-config\") pod \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.714283 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbrsc\" (UniqueName: \"kubernetes.io/projected/da52e1cb-1436-4b18-a2b6-5032d1018b0b-kube-api-access-fbrsc\") pod \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.714419 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-svc\") pod \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.714537 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-swift-storage-0\") pod \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.714639 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-sb\") pod \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.714710 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-nb\") pod \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\" (UID: \"da52e1cb-1436-4b18-a2b6-5032d1018b0b\") " Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.715552 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzrk6\" (UniqueName: \"kubernetes.io/projected/76a8192d-7bcd-4f4a-b798-907d67b3d21c-kube-api-access-kzrk6\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.715648 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327449fc-a14d-4594-b8a8-c758aeebec34-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.715704 4735 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.715762 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvbgv\" (UniqueName: \"kubernetes.io/projected/327449fc-a14d-4594-b8a8-c758aeebec34-kube-api-access-zvbgv\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.715824 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwr89\" (UniqueName: \"kubernetes.io/projected/c924b448-cd94-426c-bc25-ad0efa381777-kube-api-access-dwr89\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.715882 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.715956 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.716013 4735 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.716065 4735 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.806184 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-config-data" (OuterVolumeSpecName: "config-data") pod "c924b448-cd94-426c-bc25-ad0efa381777" (UID: "c924b448-cd94-426c-bc25-ad0efa381777"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.815232 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da52e1cb-1436-4b18-a2b6-5032d1018b0b-kube-api-access-fbrsc" (OuterVolumeSpecName: "kube-api-access-fbrsc") pod "da52e1cb-1436-4b18-a2b6-5032d1018b0b" (UID: "da52e1cb-1436-4b18-a2b6-5032d1018b0b"). InnerVolumeSpecName "kube-api-access-fbrsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.822753 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbrsc\" (UniqueName: \"kubernetes.io/projected/da52e1cb-1436-4b18-a2b6-5032d1018b0b-kube-api-access-fbrsc\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.822778 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.846202 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4r5s8" event={"ID":"c924b448-cd94-426c-bc25-ad0efa381777","Type":"ContainerDied","Data":"e4a0681ad006bfd509754a0f839b982c75ada18843c18536b9b0ddd4eb41fa57"} Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.846240 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4a0681ad006bfd509754a0f839b982c75ada18843c18536b9b0ddd4eb41fa57" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.846286 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4r5s8" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.889462 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c924b448-cd94-426c-bc25-ad0efa381777" (UID: "c924b448-cd94-426c-bc25-ad0efa381777"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.904602 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ph5qb" event={"ID":"76a8192d-7bcd-4f4a-b798-907d67b3d21c","Type":"ContainerDied","Data":"02617330b3775ed02a8aa95ad89b3c2f5d4d45f1e9a94e6337caf90f424865fd"} Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.904651 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02617330b3775ed02a8aa95ad89b3c2f5d4d45f1e9a94e6337caf90f424865fd" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.905055 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ph5qb" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.926457 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c924b448-cd94-426c-bc25-ad0efa381777-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.931532 4735 generic.go:334] "Generic (PLEG): container finished" podID="da52e1cb-1436-4b18-a2b6-5032d1018b0b" containerID="046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12" exitCode=0 Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.931607 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" event={"ID":"da52e1cb-1436-4b18-a2b6-5032d1018b0b","Type":"ContainerDied","Data":"046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12"} Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.931638 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" event={"ID":"da52e1cb-1436-4b18-a2b6-5032d1018b0b","Type":"ContainerDied","Data":"c695d6f282c9d598349eff69df6e0b8437be71602618da1fb4b65040ecf48136"} Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.931660 4735 scope.go:117] "RemoveContainer" containerID="046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.931782 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-s945p" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.944059 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76a8192d-7bcd-4f4a-b798-907d67b3d21c" (UID: "76a8192d-7bcd-4f4a-b798-907d67b3d21c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.944157 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-config-data" (OuterVolumeSpecName: "config-data") pod "327449fc-a14d-4594-b8a8-c758aeebec34" (UID: "327449fc-a14d-4594-b8a8-c758aeebec34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.949826 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "327449fc-a14d-4594-b8a8-c758aeebec34" (UID: "327449fc-a14d-4594-b8a8-c758aeebec34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.991433 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pss7t" event={"ID":"327449fc-a14d-4594-b8a8-c758aeebec34","Type":"ContainerDied","Data":"10ef2577c6b14f5922857b374e032dc0f1aa3b53d06fc744163ac5e5887b9b14"} Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.991483 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10ef2577c6b14f5922857b374e032dc0f1aa3b53d06fc744163ac5e5887b9b14" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.991536 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pss7t" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.991559 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da52e1cb-1436-4b18-a2b6-5032d1018b0b" (UID: "da52e1cb-1436-4b18-a2b6-5032d1018b0b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:33 crc kubenswrapper[4735]: I0215 20:35:33.993661 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "da52e1cb-1436-4b18-a2b6-5032d1018b0b" (UID: "da52e1cb-1436-4b18-a2b6-5032d1018b0b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.006034 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d7a00e9-2271-49b1-a74f-51f5f668995a","Type":"ContainerStarted","Data":"5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804"} Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.024342 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "da52e1cb-1436-4b18-a2b6-5032d1018b0b" (UID: "da52e1cb-1436-4b18-a2b6-5032d1018b0b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.031884 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "da52e1cb-1436-4b18-a2b6-5032d1018b0b" (UID: "da52e1cb-1436-4b18-a2b6-5032d1018b0b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.032456 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-config" (OuterVolumeSpecName: "config") pod "da52e1cb-1436-4b18-a2b6-5032d1018b0b" (UID: "da52e1cb-1436-4b18-a2b6-5032d1018b0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.035282 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.035347 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.035406 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.035461 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327449fc-a14d-4594-b8a8-c758aeebec34-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.035516 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8192d-7bcd-4f4a-b798-907d67b3d21c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.035575 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.081412 4735 scope.go:117] "RemoveContainer" containerID="355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.142981 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.143006 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da52e1cb-1436-4b18-a2b6-5032d1018b0b-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.176060 4735 scope.go:117] "RemoveContainer" containerID="046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12" Feb 15 20:35:34 crc kubenswrapper[4735]: E0215 20:35:34.177046 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12\": container with ID starting with 046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12 not found: ID does not exist" containerID="046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.177143 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12"} err="failed to get container status \"046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12\": rpc error: code = NotFound desc = could not find container \"046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12\": container with ID starting with 046410849dbd506bbc971ee1e54111e1d21005ab874f283df67c9adf90c4ce12 not found: ID does not exist" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.177218 4735 scope.go:117] "RemoveContainer" containerID="355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b" Feb 15 20:35:34 crc kubenswrapper[4735]: E0215 20:35:34.177608 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b\": container with ID starting with 355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b not found: ID does not exist" containerID="355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.177696 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b"} err="failed to get container status \"355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b\": rpc error: code = NotFound desc = could not find container \"355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b\": container with ID starting with 355fa1ed972b8919184c8a9db3dfcdfcfc7b47b92142b8c6e41aa9f8bc72c30b not found: ID does not exist" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.280838 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-s945p"] Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.289262 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-s945p"] Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.482237 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7cb4db7f5b-mrtp9"] Feb 15 20:35:34 crc kubenswrapper[4735]: E0215 20:35:34.483561 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da52e1cb-1436-4b18-a2b6-5032d1018b0b" containerName="init" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.483627 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="da52e1cb-1436-4b18-a2b6-5032d1018b0b" containerName="init" Feb 15 20:35:34 crc kubenswrapper[4735]: E0215 20:35:34.483709 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a8192d-7bcd-4f4a-b798-907d67b3d21c" containerName="barbican-db-sync" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.483760 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a8192d-7bcd-4f4a-b798-907d67b3d21c" containerName="barbican-db-sync" Feb 15 20:35:34 crc kubenswrapper[4735]: E0215 20:35:34.483857 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327449fc-a14d-4594-b8a8-c758aeebec34" containerName="placement-db-sync" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.483910 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="327449fc-a14d-4594-b8a8-c758aeebec34" containerName="placement-db-sync" Feb 15 20:35:34 crc kubenswrapper[4735]: E0215 20:35:34.484079 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c924b448-cd94-426c-bc25-ad0efa381777" containerName="keystone-bootstrap" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.484135 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c924b448-cd94-426c-bc25-ad0efa381777" containerName="keystone-bootstrap" Feb 15 20:35:34 crc kubenswrapper[4735]: E0215 20:35:34.484190 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da52e1cb-1436-4b18-a2b6-5032d1018b0b" containerName="dnsmasq-dns" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.484243 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="da52e1cb-1436-4b18-a2b6-5032d1018b0b" containerName="dnsmasq-dns" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.484467 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a8192d-7bcd-4f4a-b798-907d67b3d21c" containerName="barbican-db-sync" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.484529 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="da52e1cb-1436-4b18-a2b6-5032d1018b0b" containerName="dnsmasq-dns" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.484600 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="327449fc-a14d-4594-b8a8-c758aeebec34" containerName="placement-db-sync" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.484654 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c924b448-cd94-426c-bc25-ad0efa381777" containerName="keystone-bootstrap" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.485209 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.487128 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.487647 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.487761 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dljv4" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.489063 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.489132 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.489343 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.536519 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7cb4db7f5b-mrtp9"] Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.551110 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-fernet-keys\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.551197 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-config-data\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.551221 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-combined-ca-bundle\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.551246 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-internal-tls-certs\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.551268 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4d47\" (UniqueName: \"kubernetes.io/projected/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-kube-api-access-m4d47\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.551296 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-scripts\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.551323 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-public-tls-certs\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.551390 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-credential-keys\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.646551 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-58678bdfb8-bgfbt"] Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.649282 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652593 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-public-tls-certs\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652637 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-credential-keys\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652660 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-internal-tls-certs\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652680 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-fernet-keys\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652723 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-config-data\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652743 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-combined-ca-bundle\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652761 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-scripts\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652779 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-config-data\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652799 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-internal-tls-certs\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652813 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-combined-ca-bundle\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652833 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4d47\" (UniqueName: \"kubernetes.io/projected/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-kube-api-access-m4d47\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652859 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-scripts\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652885 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-public-tls-certs\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652912 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ec4117-00bf-4584-a22e-f760637b52ba-logs\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.652932 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmtxk\" (UniqueName: \"kubernetes.io/projected/e4ec4117-00bf-4584-a22e-f760637b52ba-kube-api-access-pmtxk\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.654155 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.654593 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.654702 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.654844 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.662486 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-config-data\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.668964 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-scripts\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.669519 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-public-tls-certs\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.670289 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-58678bdfb8-bgfbt"] Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.670961 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-tblss" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.674100 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-fernet-keys\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.676314 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-combined-ca-bundle\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.680151 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-internal-tls-certs\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.687437 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-credential-keys\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.751505 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4d47\" (UniqueName: \"kubernetes.io/projected/33778124-7c75-4b6f-a2e1-c0ac0d8eda7e-kube-api-access-m4d47\") pod \"keystone-7cb4db7f5b-mrtp9\" (UID: \"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e\") " pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.756201 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-internal-tls-certs\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.756308 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-scripts\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.756347 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-config-data\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.756373 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-combined-ca-bundle\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.756446 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ec4117-00bf-4584-a22e-f760637b52ba-logs\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.756481 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmtxk\" (UniqueName: \"kubernetes.io/projected/e4ec4117-00bf-4584-a22e-f760637b52ba-kube-api-access-pmtxk\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.756542 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-public-tls-certs\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.768004 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ec4117-00bf-4584-a22e-f760637b52ba-logs\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.768605 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-config-data\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.768655 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-public-tls-certs\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.775673 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-67c7c8d94f-trlwh"] Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.777037 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.778556 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.787091 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-b9t7b" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.787293 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.795093 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-internal-tls-certs\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.795472 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-scripts\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.813772 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.845902 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.848508 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c647fbf5b-7zs9r" podUID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.872058 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmtxk\" (UniqueName: \"kubernetes.io/projected/e4ec4117-00bf-4584-a22e-f760637b52ba-kube-api-access-pmtxk\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.872952 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-combined-ca-bundle\") pod \"placement-58678bdfb8-bgfbt\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.875650 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data-custom\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.875812 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvhhh\" (UniqueName: \"kubernetes.io/projected/a20c1cd0-5f91-41de-b20d-3711518b95ef-kube-api-access-tvhhh\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.876015 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-combined-ca-bundle\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.876143 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.876250 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a20c1cd0-5f91-41de-b20d-3711518b95ef-logs\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.877234 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.982226 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data-custom\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.982263 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvhhh\" (UniqueName: \"kubernetes.io/projected/a20c1cd0-5f91-41de-b20d-3711518b95ef-kube-api-access-tvhhh\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.982337 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-combined-ca-bundle\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.982374 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.982416 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a20c1cd0-5f91-41de-b20d-3711518b95ef-logs\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.985214 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a20c1cd0-5f91-41de-b20d-3711518b95ef-logs\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.990530 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-combined-ca-bundle\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.993900 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:34 crc kubenswrapper[4735]: I0215 20:35:34.997521 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data-custom\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.002308 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da52e1cb-1436-4b18-a2b6-5032d1018b0b" path="/var/lib/kubelet/pods/da52e1cb-1436-4b18-a2b6-5032d1018b0b/volumes" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.011058 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-684fc99c4-wchtt"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.014968 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.028962 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.043389 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvhhh\" (UniqueName: \"kubernetes.io/projected/a20c1cd0-5f91-41de-b20d-3711518b95ef-kube-api-access-tvhhh\") pod \"barbican-worker-67c7c8d94f-trlwh\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.071220 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-67c7c8d94f-trlwh"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.085075 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/904fa273-5436-4633-84f5-56c8205cd943-logs\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.085147 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-combined-ca-bundle\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.085195 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.085217 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c4lg\" (UniqueName: \"kubernetes.io/projected/904fa273-5436-4633-84f5-56c8205cd943-kube-api-access-8c4lg\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.085264 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data-custom\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.094574 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-684fc99c4-wchtt"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.108689 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6htdh" event={"ID":"ec4fe5e3-ca92-4814-8921-62117a037281","Type":"ContainerStarted","Data":"524a5b6f479153a7a9a98fa20f5ed7416bf0165d557b4e405482134294240272"} Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.128232 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zrzbw"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.129902 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.167010 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zrzbw"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.183668 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-b497d9d6c-4dqkk"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.185321 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.188630 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.188681 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.188721 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c4lg\" (UniqueName: \"kubernetes.io/projected/904fa273-5436-4633-84f5-56c8205cd943-kube-api-access-8c4lg\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.188776 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data-custom\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.188856 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpkl6\" (UniqueName: \"kubernetes.io/projected/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-kube-api-access-lpkl6\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.188873 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.188901 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-config\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.188931 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.188971 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/904fa273-5436-4633-84f5-56c8205cd943-logs\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.189026 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.189045 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-combined-ca-bundle\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.191086 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/904fa273-5436-4633-84f5-56c8205cd943-logs\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.196590 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-combined-ca-bundle\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.200244 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.204430 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data-custom\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.204906 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-b497d9d6c-4dqkk"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.219311 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c4lg\" (UniqueName: \"kubernetes.io/projected/904fa273-5436-4633-84f5-56c8205cd943-kube-api-access-8c4lg\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.236484 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-f776895b6-v9kd8"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.242725 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.251587 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data\") pod \"barbican-keystone-listener-684fc99c4-wchtt\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291198 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-config-data-custom\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291272 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47t7x\" (UniqueName: \"kubernetes.io/projected/8c4ffab0-9e1c-4ccb-841d-54c812217899-kube-api-access-47t7x\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291309 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpkl6\" (UniqueName: \"kubernetes.io/projected/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-kube-api-access-lpkl6\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291334 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291367 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-config\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291394 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291423 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c4ffab0-9e1c-4ccb-841d-54c812217899-combined-ca-bundle\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291481 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c4ffab0-9e1c-4ccb-841d-54c812217899-config-data\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291516 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c4ffab0-9e1c-4ccb-841d-54c812217899-logs\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291537 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c4ffab0-9e1c-4ccb-841d-54c812217899-config-data-custom\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291563 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291596 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291620 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-logs\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291654 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-config-data\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291702 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgb5m\" (UniqueName: \"kubernetes.io/projected/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-kube-api-access-qgb5m\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291730 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-combined-ca-bundle\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291974 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-6htdh" podStartSLOduration=4.958332618 podStartE2EDuration="51.291938458s" podCreationTimestamp="2026-02-15 20:34:44 +0000 UTC" firstStartedPulling="2026-02-15 20:34:46.895094546 +0000 UTC m=+1094.761110169" lastFinishedPulling="2026-02-15 20:35:33.228700386 +0000 UTC m=+1141.094716009" observedRunningTime="2026-02-15 20:35:35.142360182 +0000 UTC m=+1143.008375805" watchObservedRunningTime="2026-02-15 20:35:35.291938458 +0000 UTC m=+1143.157954081" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.291099 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-f776895b6-v9kd8"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.320094 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.322994 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-config\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.324527 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.325101 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.336830 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.358278 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.373768 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpkl6\" (UniqueName: \"kubernetes.io/projected/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-kube-api-access-lpkl6\") pod \"dnsmasq-dns-848cf88cfc-zrzbw\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.419114 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.421759 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-config-data-custom\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.421820 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47t7x\" (UniqueName: \"kubernetes.io/projected/8c4ffab0-9e1c-4ccb-841d-54c812217899-kube-api-access-47t7x\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.421868 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c4ffab0-9e1c-4ccb-841d-54c812217899-combined-ca-bundle\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.421915 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c4ffab0-9e1c-4ccb-841d-54c812217899-config-data\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.421958 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c4ffab0-9e1c-4ccb-841d-54c812217899-logs\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.421974 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c4ffab0-9e1c-4ccb-841d-54c812217899-config-data-custom\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.422000 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-logs\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.422026 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-config-data\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.422069 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgb5m\" (UniqueName: \"kubernetes.io/projected/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-kube-api-access-qgb5m\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.422086 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-combined-ca-bundle\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.423141 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c4ffab0-9e1c-4ccb-841d-54c812217899-logs\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.425777 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-logs\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.429962 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-combined-ca-bundle\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.442688 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-config-data\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.448963 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c4ffab0-9e1c-4ccb-841d-54c812217899-config-data-custom\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.449231 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c4ffab0-9e1c-4ccb-841d-54c812217899-config-data\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.459897 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-config-data-custom\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.469571 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c4ffab0-9e1c-4ccb-841d-54c812217899-combined-ca-bundle\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.471168 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgb5m\" (UniqueName: \"kubernetes.io/projected/29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2-kube-api-access-qgb5m\") pod \"barbican-worker-b497d9d6c-4dqkk\" (UID: \"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2\") " pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.502392 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6c5c4f5b9b-sg99c"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.504570 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.518809 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47t7x\" (UniqueName: \"kubernetes.io/projected/8c4ffab0-9e1c-4ccb-841d-54c812217899-kube-api-access-47t7x\") pod \"barbican-keystone-listener-f776895b6-v9kd8\" (UID: \"8c4ffab0-9e1c-4ccb-841d-54c812217899\") " pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.523608 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-logs\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.523639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-combined-ca-bundle\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.523662 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-internal-tls-certs\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.523713 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-config-data\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.523763 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-public-tls-certs\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.523860 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-scripts\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.523917 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ttj5\" (UniqueName: \"kubernetes.io/projected/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-kube-api-access-5ttj5\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.561505 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6c5c4f5b9b-sg99c"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.576374 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7c7df566b8-mgngk"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.577794 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.581692 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.615776 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c7df566b8-mgngk"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.626379 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-logs\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.627725 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-combined-ca-bundle\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.628886 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-internal-tls-certs\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.629110 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-config-data\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.629823 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-public-tls-certs\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.630381 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-scripts\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.630715 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ttj5\" (UniqueName: \"kubernetes.io/projected/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-kube-api-access-5ttj5\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.628258 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-logs\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.642638 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-combined-ca-bundle\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.646387 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-scripts\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.651886 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-config-data\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.652557 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-internal-tls-certs\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.666416 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-public-tls-certs\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.672417 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ttj5\" (UniqueName: \"kubernetes.io/projected/da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f-kube-api-access-5ttj5\") pod \"placement-6c5c4f5b9b-sg99c\" (UID: \"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f\") " pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.720965 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b497d9d6c-4dqkk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.732860 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.732910 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-combined-ca-bundle\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.732961 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5391fc9-2f13-446e-9b98-8f453c23393f-logs\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.733011 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data-custom\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.733066 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lscs\" (UniqueName: \"kubernetes.io/projected/c5391fc9-2f13-446e-9b98-8f453c23393f-kube-api-access-8lscs\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.748167 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.823484 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7cb4db7f5b-mrtp9"] Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.843711 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data-custom\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.843797 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lscs\" (UniqueName: \"kubernetes.io/projected/c5391fc9-2f13-446e-9b98-8f453c23393f-kube-api-access-8lscs\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.843836 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.843870 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-combined-ca-bundle\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.843902 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5391fc9-2f13-446e-9b98-8f453c23393f-logs\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.846655 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5391fc9-2f13-446e-9b98-8f453c23393f-logs\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.857187 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-combined-ca-bundle\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.861470 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.866298 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.866709 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data-custom\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.875412 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lscs\" (UniqueName: \"kubernetes.io/projected/c5391fc9-2f13-446e-9b98-8f453c23393f-kube-api-access-8lscs\") pod \"barbican-api-7c7df566b8-mgngk\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.904599 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:35 crc kubenswrapper[4735]: I0215 20:35:35.996320 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-58678bdfb8-bgfbt"] Feb 15 20:35:36 crc kubenswrapper[4735]: I0215 20:35:36.167578 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7cb4db7f5b-mrtp9" event={"ID":"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e","Type":"ContainerStarted","Data":"8267d08a88e8abcf8594c45d979b75ee322ec4d9272c5aa854f790f5a49a3d57"} Feb 15 20:35:36 crc kubenswrapper[4735]: I0215 20:35:36.216601 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58678bdfb8-bgfbt" event={"ID":"e4ec4117-00bf-4584-a22e-f760637b52ba","Type":"ContainerStarted","Data":"2b35d9cd5da1bab6bf0fe1d5421feceb001aea15433be545c81b8185a6f5a014"} Feb 15 20:35:36 crc kubenswrapper[4735]: I0215 20:35:36.804846 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zrzbw"] Feb 15 20:35:36 crc kubenswrapper[4735]: I0215 20:35:36.868165 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-67c7c8d94f-trlwh"] Feb 15 20:35:36 crc kubenswrapper[4735]: I0215 20:35:36.957507 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-684fc99c4-wchtt"] Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.285240 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" event={"ID":"c49c8247-c8d6-4f83-9b3c-565cb6fefe51","Type":"ContainerStarted","Data":"f6718d23c3977119b2b7cd3a510abd92955a136713ea763478c6ed74a3b5def5"} Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.291019 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67c7c8d94f-trlwh" event={"ID":"a20c1cd0-5f91-41de-b20d-3711518b95ef","Type":"ContainerStarted","Data":"c340ab8e64181fba0ff6704d6f0e337cde1acf656755cf8cb751b453aabbe204"} Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.311198 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7cb4db7f5b-mrtp9" event={"ID":"33778124-7c75-4b6f-a2e1-c0ac0d8eda7e","Type":"ContainerStarted","Data":"bb51b17dc34451310e4eafd73fcca805d23fdc167cd49ac88e93ac54a4c3c85b"} Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.311509 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.314791 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" event={"ID":"904fa273-5436-4633-84f5-56c8205cd943","Type":"ContainerStarted","Data":"fd36fac619467d9278527ca97669feaa20216178123d0981b2c4364d32cf44fc"} Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.325560 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58678bdfb8-bgfbt" event={"ID":"e4ec4117-00bf-4584-a22e-f760637b52ba","Type":"ContainerStarted","Data":"84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f"} Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.338587 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7cb4db7f5b-mrtp9" podStartSLOduration=3.338576172 podStartE2EDuration="3.338576172s" podCreationTimestamp="2026-02-15 20:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:37.33364156 +0000 UTC m=+1145.199657183" watchObservedRunningTime="2026-02-15 20:35:37.338576172 +0000 UTC m=+1145.204591795" Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.377037 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-f776895b6-v9kd8"] Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.392561 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6c5c4f5b9b-sg99c"] Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.406831 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c7df566b8-mgngk"] Feb 15 20:35:37 crc kubenswrapper[4735]: I0215 20:35:37.468357 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-b497d9d6c-4dqkk"] Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.355693 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58678bdfb8-bgfbt" event={"ID":"e4ec4117-00bf-4584-a22e-f760637b52ba","Type":"ContainerStarted","Data":"991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226"} Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.356389 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.356405 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.357643 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b497d9d6c-4dqkk" event={"ID":"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2","Type":"ContainerStarted","Data":"a0b198a76289263a01ffeeff5d47217813bb0caaf8fcf58393fa875d774d1492"} Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.359607 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c7df566b8-mgngk" event={"ID":"c5391fc9-2f13-446e-9b98-8f453c23393f","Type":"ContainerStarted","Data":"6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5"} Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.359633 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c7df566b8-mgngk" event={"ID":"c5391fc9-2f13-446e-9b98-8f453c23393f","Type":"ContainerStarted","Data":"13cd91783b5dd561d1cd0c360c9e3a96028e26b686a692190518164fa8b6d4e1"} Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.369487 4735 generic.go:334] "Generic (PLEG): container finished" podID="c49c8247-c8d6-4f83-9b3c-565cb6fefe51" containerID="0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263" exitCode=0 Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.369544 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" event={"ID":"c49c8247-c8d6-4f83-9b3c-565cb6fefe51","Type":"ContainerDied","Data":"0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263"} Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.377505 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" event={"ID":"8c4ffab0-9e1c-4ccb-841d-54c812217899","Type":"ContainerStarted","Data":"b41a4bb82a83be831654e3f44abc0c2004adbf34712156728b3e37b998f99f6a"} Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.379330 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-58678bdfb8-bgfbt" podStartSLOduration=4.37931967 podStartE2EDuration="4.37931967s" podCreationTimestamp="2026-02-15 20:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:38.375335362 +0000 UTC m=+1146.241350985" watchObservedRunningTime="2026-02-15 20:35:38.37931967 +0000 UTC m=+1146.245335293" Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.388978 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c5c4f5b9b-sg99c" event={"ID":"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f","Type":"ContainerStarted","Data":"8a32de398a6c82020339b5f621fd3749aa1c1e1bb5b51ee29e655a184a20b50c"} Feb 15 20:35:38 crc kubenswrapper[4735]: I0215 20:35:38.389011 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c5c4f5b9b-sg99c" event={"ID":"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f","Type":"ContainerStarted","Data":"148b928405b93a34abdd02303ba0448f0e6587e8a7925c506c485bc9b8ddc512"} Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.428274 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c7df566b8-mgngk" event={"ID":"c5391fc9-2f13-446e-9b98-8f453c23393f","Type":"ContainerStarted","Data":"5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e"} Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.428665 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.428700 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.450566 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7d4fd7fbc4-nt694"] Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.453040 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.469252 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.469654 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.525397 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7c7df566b8-mgngk" podStartSLOduration=4.525381418 podStartE2EDuration="4.525381418s" podCreationTimestamp="2026-02-15 20:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:39.485580375 +0000 UTC m=+1147.351595998" watchObservedRunningTime="2026-02-15 20:35:39.525381418 +0000 UTC m=+1147.391397041" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.525989 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" event={"ID":"c49c8247-c8d6-4f83-9b3c-565cb6fefe51","Type":"ContainerStarted","Data":"f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15"} Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.526087 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d4fd7fbc4-nt694"] Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.526154 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.534917 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c5c4f5b9b-sg99c" event={"ID":"da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f","Type":"ContainerStarted","Data":"e247586b34195f1ce25e7a664761ab8ec1ee15203b927455214c09cc6e719282"} Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.535091 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.536173 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.568367 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-combined-ca-bundle\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.568634 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-config-data-custom\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.568766 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxcn4\" (UniqueName: \"kubernetes.io/projected/c0df081d-fe76-415d-b26f-c67e2c97284f-kube-api-access-nxcn4\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.568855 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0df081d-fe76-415d-b26f-c67e2c97284f-logs\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.568952 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-config-data\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.569068 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-internal-tls-certs\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.569197 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-public-tls-certs\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.589613 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6c5c4f5b9b-sg99c" podStartSLOduration=4.589596141 podStartE2EDuration="4.589596141s" podCreationTimestamp="2026-02-15 20:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:39.570465544 +0000 UTC m=+1147.436481167" watchObservedRunningTime="2026-02-15 20:35:39.589596141 +0000 UTC m=+1147.455611764" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.648087 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" podStartSLOduration=5.648070818 podStartE2EDuration="5.648070818s" podCreationTimestamp="2026-02-15 20:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:39.635634682 +0000 UTC m=+1147.501650305" watchObservedRunningTime="2026-02-15 20:35:39.648070818 +0000 UTC m=+1147.514086441" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.674087 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-internal-tls-certs\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.674230 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-public-tls-certs\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.674340 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-combined-ca-bundle\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.674457 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-config-data-custom\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.674560 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxcn4\" (UniqueName: \"kubernetes.io/projected/c0df081d-fe76-415d-b26f-c67e2c97284f-kube-api-access-nxcn4\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.674583 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0df081d-fe76-415d-b26f-c67e2c97284f-logs\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.674610 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-config-data\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.687931 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-combined-ca-bundle\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.689424 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-internal-tls-certs\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.680802 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0df081d-fe76-415d-b26f-c67e2c97284f-logs\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.705650 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-config-data-custom\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.709332 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-config-data\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.719710 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0df081d-fe76-415d-b26f-c67e2c97284f-public-tls-certs\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.746712 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxcn4\" (UniqueName: \"kubernetes.io/projected/c0df081d-fe76-415d-b26f-c67e2c97284f-kube-api-access-nxcn4\") pod \"barbican-api-7d4fd7fbc4-nt694\" (UID: \"c0df081d-fe76-415d-b26f-c67e2c97284f\") " pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:39 crc kubenswrapper[4735]: I0215 20:35:39.789273 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:40 crc kubenswrapper[4735]: I0215 20:35:40.792871 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 15 20:35:42 crc kubenswrapper[4735]: I0215 20:35:42.454203 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d4fd7fbc4-nt694"] Feb 15 20:35:42 crc kubenswrapper[4735]: I0215 20:35:42.588079 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d4fd7fbc4-nt694" event={"ID":"c0df081d-fe76-415d-b26f-c67e2c97284f","Type":"ContainerStarted","Data":"0b8d4fd451a23e44e4c82edde505f3e4d174d657becdef2b96bdae39847a9319"} Feb 15 20:35:42 crc kubenswrapper[4735]: I0215 20:35:42.596277 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" event={"ID":"8c4ffab0-9e1c-4ccb-841d-54c812217899","Type":"ContainerStarted","Data":"a5675def7d5ae3f0d0db4b3fe726637e1c9e79759f86422a7510138524faa65a"} Feb 15 20:35:42 crc kubenswrapper[4735]: I0215 20:35:42.623915 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67c7c8d94f-trlwh" event={"ID":"a20c1cd0-5f91-41de-b20d-3711518b95ef","Type":"ContainerStarted","Data":"eefa0ed78309417ca1db111a3d8580f1c3369b7be5c327884253132a62d51333"} Feb 15 20:35:42 crc kubenswrapper[4735]: I0215 20:35:42.626110 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" event={"ID":"904fa273-5436-4633-84f5-56c8205cd943","Type":"ContainerStarted","Data":"4a8cd143180cf340ef0b1a459c8443020429f5971a4b804256913003887dc484"} Feb 15 20:35:42 crc kubenswrapper[4735]: I0215 20:35:42.627286 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b497d9d6c-4dqkk" event={"ID":"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2","Type":"ContainerStarted","Data":"6f55e36c617dd42aadf8adebdf99028920be83c5f95ee04fe1fe78ce51ca1ec9"} Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.642218 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d4fd7fbc4-nt694" event={"ID":"c0df081d-fe76-415d-b26f-c67e2c97284f","Type":"ContainerStarted","Data":"fbbb0cd34e37cc5f6974bb495bc88f6d470e090d5ff385e89bf8b5f8cccbffd7"} Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.648176 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" event={"ID":"8c4ffab0-9e1c-4ccb-841d-54c812217899","Type":"ContainerStarted","Data":"ffc75ca8709db96b3f18a52db2d36e521c82c3c813c30749f6fce52c61bbc276"} Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.651472 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67c7c8d94f-trlwh" event={"ID":"a20c1cd0-5f91-41de-b20d-3711518b95ef","Type":"ContainerStarted","Data":"259bea590336172fa97888d69b7055d1c352e66a801b2b09b71e1977436fca5a"} Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.656082 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" event={"ID":"904fa273-5436-4633-84f5-56c8205cd943","Type":"ContainerStarted","Data":"151537ee184a3d9639d7aad06d1ab88b4be09b9b675a80f059d210450a6c0d98"} Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.662477 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b497d9d6c-4dqkk" event={"ID":"29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2","Type":"ContainerStarted","Data":"cf1aa1b5509b0cb4bbe8289a9ba81b4c4f0c315090e34aa6e6173288e7b2c06f"} Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.688829 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-f776895b6-v9kd8" podStartSLOduration=5.187910953 podStartE2EDuration="9.688812998s" podCreationTimestamp="2026-02-15 20:35:34 +0000 UTC" firstStartedPulling="2026-02-15 20:35:37.433088422 +0000 UTC m=+1145.299104045" lastFinishedPulling="2026-02-15 20:35:41.933990467 +0000 UTC m=+1149.800006090" observedRunningTime="2026-02-15 20:35:43.683809473 +0000 UTC m=+1151.549825096" watchObservedRunningTime="2026-02-15 20:35:43.688812998 +0000 UTC m=+1151.554828621" Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.751996 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-b497d9d6c-4dqkk" podStartSLOduration=5.38757926 podStartE2EDuration="9.751976052s" podCreationTimestamp="2026-02-15 20:35:34 +0000 UTC" firstStartedPulling="2026-02-15 20:35:37.570102659 +0000 UTC m=+1145.436118282" lastFinishedPulling="2026-02-15 20:35:41.934499451 +0000 UTC m=+1149.800515074" observedRunningTime="2026-02-15 20:35:43.743057382 +0000 UTC m=+1151.609073015" watchObservedRunningTime="2026-02-15 20:35:43.751976052 +0000 UTC m=+1151.617991675" Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.792305 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" podStartSLOduration=4.859280928 podStartE2EDuration="9.79228854s" podCreationTimestamp="2026-02-15 20:35:34 +0000 UTC" firstStartedPulling="2026-02-15 20:35:37.001980422 +0000 UTC m=+1144.867996045" lastFinishedPulling="2026-02-15 20:35:41.934988034 +0000 UTC m=+1149.801003657" observedRunningTime="2026-02-15 20:35:43.78044521 +0000 UTC m=+1151.646460833" watchObservedRunningTime="2026-02-15 20:35:43.79228854 +0000 UTC m=+1151.658304163" Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.849305 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-67c7c8d94f-trlwh" podStartSLOduration=4.8235226319999995 podStartE2EDuration="9.849292167s" podCreationTimestamp="2026-02-15 20:35:34 +0000 UTC" firstStartedPulling="2026-02-15 20:35:36.90704518 +0000 UTC m=+1144.773060803" lastFinishedPulling="2026-02-15 20:35:41.932814715 +0000 UTC m=+1149.798830338" observedRunningTime="2026-02-15 20:35:43.837291944 +0000 UTC m=+1151.703307557" watchObservedRunningTime="2026-02-15 20:35:43.849292167 +0000 UTC m=+1151.715307780" Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.849856 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-67c7c8d94f-trlwh"] Feb 15 20:35:43 crc kubenswrapper[4735]: I0215 20:35:43.876493 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-684fc99c4-wchtt"] Feb 15 20:35:44 crc kubenswrapper[4735]: I0215 20:35:44.679548 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d4fd7fbc4-nt694" event={"ID":"c0df081d-fe76-415d-b26f-c67e2c97284f","Type":"ContainerStarted","Data":"37e55198531ca2dccc47187ad27969381293cd17578ca484651c4076f1d07f8e"} Feb 15 20:35:44 crc kubenswrapper[4735]: I0215 20:35:44.687124 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:44 crc kubenswrapper[4735]: I0215 20:35:44.687199 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:44 crc kubenswrapper[4735]: I0215 20:35:44.737906 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7d4fd7fbc4-nt694" podStartSLOduration=5.73788778 podStartE2EDuration="5.73788778s" podCreationTimestamp="2026-02-15 20:35:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:44.73345081 +0000 UTC m=+1152.599466443" watchObservedRunningTime="2026-02-15 20:35:44.73788778 +0000 UTC m=+1152.603903403" Feb 15 20:35:44 crc kubenswrapper[4735]: I0215 20:35:44.777969 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:35:44 crc kubenswrapper[4735]: I0215 20:35:44.847788 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c647fbf5b-7zs9r" podUID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 15 20:35:45 crc kubenswrapper[4735]: I0215 20:35:45.420091 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:35:45 crc kubenswrapper[4735]: I0215 20:35:45.488018 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8ggxh"] Feb 15 20:35:45 crc kubenswrapper[4735]: I0215 20:35:45.488223 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" podUID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerName="dnsmasq-dns" containerID="cri-o://dfe66441eaccfa6b08bdac1ca3e8b4ad52dc69c9543ac02fe7622c5f3c0ce6b0" gracePeriod=10 Feb 15 20:35:45 crc kubenswrapper[4735]: I0215 20:35:45.708820 4735 generic.go:334] "Generic (PLEG): container finished" podID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerID="dfe66441eaccfa6b08bdac1ca3e8b4ad52dc69c9543ac02fe7622c5f3c0ce6b0" exitCode=0 Feb 15 20:35:45 crc kubenswrapper[4735]: I0215 20:35:45.709392 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" podUID="904fa273-5436-4633-84f5-56c8205cd943" containerName="barbican-keystone-listener-log" containerID="cri-o://4a8cd143180cf340ef0b1a459c8443020429f5971a4b804256913003887dc484" gracePeriod=30 Feb 15 20:35:45 crc kubenswrapper[4735]: I0215 20:35:45.709542 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-67c7c8d94f-trlwh" podUID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerName="barbican-worker-log" containerID="cri-o://eefa0ed78309417ca1db111a3d8580f1c3369b7be5c327884253132a62d51333" gracePeriod=30 Feb 15 20:35:45 crc kubenswrapper[4735]: I0215 20:35:45.709626 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-67c7c8d94f-trlwh" podUID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerName="barbican-worker" containerID="cri-o://259bea590336172fa97888d69b7055d1c352e66a801b2b09b71e1977436fca5a" gracePeriod=30 Feb 15 20:35:45 crc kubenswrapper[4735]: I0215 20:35:45.709682 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" podUID="904fa273-5436-4633-84f5-56c8205cd943" containerName="barbican-keystone-listener" containerID="cri-o://151537ee184a3d9639d7aad06d1ab88b4be09b9b675a80f059d210450a6c0d98" gracePeriod=30 Feb 15 20:35:45 crc kubenswrapper[4735]: I0215 20:35:45.709783 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" event={"ID":"ed582df2-f4b9-4c7d-9f67-695b388d649d","Type":"ContainerDied","Data":"dfe66441eaccfa6b08bdac1ca3e8b4ad52dc69c9543ac02fe7622c5f3c0ce6b0"} Feb 15 20:35:46 crc kubenswrapper[4735]: I0215 20:35:46.724577 4735 generic.go:334] "Generic (PLEG): container finished" podID="904fa273-5436-4633-84f5-56c8205cd943" containerID="4a8cd143180cf340ef0b1a459c8443020429f5971a4b804256913003887dc484" exitCode=143 Feb 15 20:35:46 crc kubenswrapper[4735]: I0215 20:35:46.724679 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" event={"ID":"904fa273-5436-4633-84f5-56c8205cd943","Type":"ContainerDied","Data":"4a8cd143180cf340ef0b1a459c8443020429f5971a4b804256913003887dc484"} Feb 15 20:35:46 crc kubenswrapper[4735]: I0215 20:35:46.729740 4735 generic.go:334] "Generic (PLEG): container finished" podID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerID="eefa0ed78309417ca1db111a3d8580f1c3369b7be5c327884253132a62d51333" exitCode=143 Feb 15 20:35:46 crc kubenswrapper[4735]: I0215 20:35:46.730597 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67c7c8d94f-trlwh" event={"ID":"a20c1cd0-5f91-41de-b20d-3711518b95ef","Type":"ContainerDied","Data":"eefa0ed78309417ca1db111a3d8580f1c3369b7be5c327884253132a62d51333"} Feb 15 20:35:47 crc kubenswrapper[4735]: I0215 20:35:47.740789 4735 generic.go:334] "Generic (PLEG): container finished" podID="ec4fe5e3-ca92-4814-8921-62117a037281" containerID="524a5b6f479153a7a9a98fa20f5ed7416bf0165d557b4e405482134294240272" exitCode=0 Feb 15 20:35:47 crc kubenswrapper[4735]: I0215 20:35:47.740835 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6htdh" event={"ID":"ec4fe5e3-ca92-4814-8921-62117a037281","Type":"ContainerDied","Data":"524a5b6f479153a7a9a98fa20f5ed7416bf0165d557b4e405482134294240272"} Feb 15 20:35:47 crc kubenswrapper[4735]: I0215 20:35:47.829713 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" podUID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: connect: connection refused" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.001055 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.253606 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-557bd9599f-5lqht"] Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.253926 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-557bd9599f-5lqht" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerName="neutron-api" containerID="cri-o://7a7b24352b641cb77fad01a59540c8459a7c1234dcf8766c7319d7e46daa5006" gracePeriod=30 Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.254384 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-557bd9599f-5lqht" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerName="neutron-httpd" containerID="cri-o://eb9e7b1d420557821175350c51326aa947abdf4c0c45729785ed5d87a383f30c" gracePeriod=30 Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.359524 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-fdf549c7-xpft7"] Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.361431 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.384493 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fdf549c7-xpft7"] Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.391274 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.392544 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-public-tls-certs\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.392574 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-config\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.392610 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwm6x\" (UniqueName: \"kubernetes.io/projected/33cc02bd-6580-4d99-9c26-5f0c222e6461-kube-api-access-vwm6x\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.392637 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-combined-ca-bundle\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.392691 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-internal-tls-certs\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.392727 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-ovndb-tls-certs\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.392741 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-httpd-config\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.494462 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-internal-tls-certs\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.494797 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-ovndb-tls-certs\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.494921 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-httpd-config\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.495589 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-public-tls-certs\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.495693 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-config\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.495821 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwm6x\" (UniqueName: \"kubernetes.io/projected/33cc02bd-6580-4d99-9c26-5f0c222e6461-kube-api-access-vwm6x\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.495912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-combined-ca-bundle\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.506653 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-httpd-config\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.506683 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-public-tls-certs\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.507315 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-config\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.507630 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-combined-ca-bundle\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.518796 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-ovndb-tls-certs\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.520380 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cc02bd-6580-4d99-9c26-5f0c222e6461-internal-tls-certs\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.535738 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwm6x\" (UniqueName: \"kubernetes.io/projected/33cc02bd-6580-4d99-9c26-5f0c222e6461-kube-api-access-vwm6x\") pod \"neutron-fdf549c7-xpft7\" (UID: \"33cc02bd-6580-4d99-9c26-5f0c222e6461\") " pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:48 crc kubenswrapper[4735]: I0215 20:35:48.692578 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:49 crc kubenswrapper[4735]: I0215 20:35:49.680439 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:35:49 crc kubenswrapper[4735]: I0215 20:35:49.680684 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:35:49 crc kubenswrapper[4735]: I0215 20:35:49.766064 4735 generic.go:334] "Generic (PLEG): container finished" podID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerID="eb9e7b1d420557821175350c51326aa947abdf4c0c45729785ed5d87a383f30c" exitCode=0 Feb 15 20:35:49 crc kubenswrapper[4735]: I0215 20:35:49.766104 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557bd9599f-5lqht" event={"ID":"2ef1315b-689a-4bab-9e16-0f767cd22170","Type":"ContainerDied","Data":"eb9e7b1d420557821175350c51326aa947abdf4c0c45729785ed5d87a383f30c"} Feb 15 20:35:49 crc kubenswrapper[4735]: I0215 20:35:49.873077 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:49 crc kubenswrapper[4735]: I0215 20:35:49.948125 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7c7df566b8-mgngk" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 15 20:35:50 crc kubenswrapper[4735]: I0215 20:35:50.050322 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:50 crc kubenswrapper[4735]: I0215 20:35:50.656409 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-557bd9599f-5lqht" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.157:9696/\": dial tcp 10.217.0.157:9696: connect: connection refused" Feb 15 20:35:50 crc kubenswrapper[4735]: I0215 20:35:50.783576 4735 generic.go:334] "Generic (PLEG): container finished" podID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerID="10a22afe76edb65bfb324a00acf60c68c9864393f0fe03999f09757222e33183" exitCode=137 Feb 15 20:35:50 crc kubenswrapper[4735]: I0215 20:35:50.783615 4735 generic.go:334] "Generic (PLEG): container finished" podID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerID="81739c66d31bfab9b2a5e97080b0eccbabd62978dd60f71980b7bfc6c8bc8839" exitCode=137 Feb 15 20:35:50 crc kubenswrapper[4735]: I0215 20:35:50.783621 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f7d59545-hbprm" event={"ID":"1be12817-7c08-4d66-938d-98ce9f79e4fd","Type":"ContainerDied","Data":"10a22afe76edb65bfb324a00acf60c68c9864393f0fe03999f09757222e33183"} Feb 15 20:35:50 crc kubenswrapper[4735]: I0215 20:35:50.783681 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f7d59545-hbprm" event={"ID":"1be12817-7c08-4d66-938d-98ce9f79e4fd","Type":"ContainerDied","Data":"81739c66d31bfab9b2a5e97080b0eccbabd62978dd60f71980b7bfc6c8bc8839"} Feb 15 20:35:50 crc kubenswrapper[4735]: I0215 20:35:50.785680 4735 generic.go:334] "Generic (PLEG): container finished" podID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerID="7a7b24352b641cb77fad01a59540c8459a7c1234dcf8766c7319d7e46daa5006" exitCode=0 Feb 15 20:35:50 crc kubenswrapper[4735]: I0215 20:35:50.785764 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557bd9599f-5lqht" event={"ID":"2ef1315b-689a-4bab-9e16-0f767cd22170","Type":"ContainerDied","Data":"7a7b24352b641cb77fad01a59540c8459a7c1234dcf8766c7319d7e46daa5006"} Feb 15 20:35:51 crc kubenswrapper[4735]: W0215 20:35:51.712101 4735 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc49c8247_c8d6_4f83_9b3c_565cb6fefe51.slice/crio-conmon-0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc49c8247_c8d6_4f83_9b3c_565cb6fefe51.slice/crio-conmon-0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263.scope: no such file or directory Feb 15 20:35:51 crc kubenswrapper[4735]: W0215 20:35:51.712359 4735 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc49c8247_c8d6_4f83_9b3c_565cb6fefe51.slice/crio-0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc49c8247_c8d6_4f83_9b3c_565cb6fefe51.slice/crio-0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263.scope: no such file or directory Feb 15 20:35:51 crc kubenswrapper[4735]: W0215 20:35:51.715409 4735 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod904fa273_5436_4633_84f5_56c8205cd943.slice/crio-conmon-4a8cd143180cf340ef0b1a459c8443020429f5971a4b804256913003887dc484.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod904fa273_5436_4633_84f5_56c8205cd943.slice/crio-conmon-4a8cd143180cf340ef0b1a459c8443020429f5971a4b804256913003887dc484.scope: no such file or directory Feb 15 20:35:51 crc kubenswrapper[4735]: W0215 20:35:51.715436 4735 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda20c1cd0_5f91_41de_b20d_3711518b95ef.slice/crio-conmon-eefa0ed78309417ca1db111a3d8580f1c3369b7be5c327884253132a62d51333.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda20c1cd0_5f91_41de_b20d_3711518b95ef.slice/crio-conmon-eefa0ed78309417ca1db111a3d8580f1c3369b7be5c327884253132a62d51333.scope: no such file or directory Feb 15 20:35:51 crc kubenswrapper[4735]: W0215 20:35:51.715456 4735 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod904fa273_5436_4633_84f5_56c8205cd943.slice/crio-4a8cd143180cf340ef0b1a459c8443020429f5971a4b804256913003887dc484.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod904fa273_5436_4633_84f5_56c8205cd943.slice/crio-4a8cd143180cf340ef0b1a459c8443020429f5971a4b804256913003887dc484.scope: no such file or directory Feb 15 20:35:51 crc kubenswrapper[4735]: W0215 20:35:51.715505 4735 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda20c1cd0_5f91_41de_b20d_3711518b95ef.slice/crio-eefa0ed78309417ca1db111a3d8580f1c3369b7be5c327884253132a62d51333.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda20c1cd0_5f91_41de_b20d_3711518b95ef.slice/crio-eefa0ed78309417ca1db111a3d8580f1c3369b7be5c327884253132a62d51333.scope: no such file or directory Feb 15 20:35:51 crc kubenswrapper[4735]: I0215 20:35:51.801763 4735 generic.go:334] "Generic (PLEG): container finished" podID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerID="363de19826e86b82ff64c234ca3ec5e5b9bfa1a8265f4b25fb7531f400e00c7c" exitCode=137 Feb 15 20:35:51 crc kubenswrapper[4735]: I0215 20:35:51.801798 4735 generic.go:334] "Generic (PLEG): container finished" podID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerID="f295535d4a74689debfd09b839f03d0ba86e654fdda24a7807dcd72de4f9eb23" exitCode=137 Feb 15 20:35:51 crc kubenswrapper[4735]: I0215 20:35:51.801834 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74dd7854c7-t2t6r" event={"ID":"faf23e04-da47-4121-9e55-d003d8d6dcfe","Type":"ContainerDied","Data":"363de19826e86b82ff64c234ca3ec5e5b9bfa1a8265f4b25fb7531f400e00c7c"} Feb 15 20:35:51 crc kubenswrapper[4735]: I0215 20:35:51.801859 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74dd7854c7-t2t6r" event={"ID":"faf23e04-da47-4121-9e55-d003d8d6dcfe","Type":"ContainerDied","Data":"f295535d4a74689debfd09b839f03d0ba86e654fdda24a7807dcd72de4f9eb23"} Feb 15 20:35:51 crc kubenswrapper[4735]: I0215 20:35:51.814358 4735 generic.go:334] "Generic (PLEG): container finished" podID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerID="5c3c8f494e812ff4a8c0aee1120fe0bd32c14ec07a9effb34a54ef3370575760" exitCode=137 Feb 15 20:35:51 crc kubenswrapper[4735]: I0215 20:35:51.814388 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6898675dc5-fh44t" event={"ID":"a6f89bfa-6fb3-40d0-9741-cfddddb9a685","Type":"ContainerDied","Data":"5c3c8f494e812ff4a8c0aee1120fe0bd32c14ec07a9effb34a54ef3370575760"} Feb 15 20:35:51 crc kubenswrapper[4735]: E0215 20:35:51.918141 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76a8192d_7bcd_4f4a_b798_907d67b3d21c.slice/crio-02617330b3775ed02a8aa95ad89b3c2f5d4d45f1e9a94e6337caf90f424865fd\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec4fe5e3_ca92_4814_8921_62117a037281.slice/crio-conmon-524a5b6f479153a7a9a98fa20f5ed7416bf0165d557b4e405482134294240272.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec4fe5e3_ca92_4814_8921_62117a037281.slice/crio-524a5b6f479153a7a9a98fa20f5ed7416bf0165d557b4e405482134294240272.scope\": RecentStats: unable to find data in memory cache]" Feb 15 20:35:52 crc kubenswrapper[4735]: I0215 20:35:52.641614 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:52 crc kubenswrapper[4735]: I0215 20:35:52.827906 4735 generic.go:334] "Generic (PLEG): container finished" podID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerID="a3e86739b52ad23dbac1a7f4cfcf2a070a9b0f25ee1f6dad3b0794fcaa858d61" exitCode=137 Feb 15 20:35:52 crc kubenswrapper[4735]: I0215 20:35:52.828024 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6898675dc5-fh44t" event={"ID":"a6f89bfa-6fb3-40d0-9741-cfddddb9a685","Type":"ContainerDied","Data":"a3e86739b52ad23dbac1a7f4cfcf2a070a9b0f25ee1f6dad3b0794fcaa858d61"} Feb 15 20:35:52 crc kubenswrapper[4735]: I0215 20:35:52.830595 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" podUID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: connect: connection refused" Feb 15 20:35:52 crc kubenswrapper[4735]: I0215 20:35:52.902425 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d4fd7fbc4-nt694" Feb 15 20:35:52 crc kubenswrapper[4735]: I0215 20:35:52.966570 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c7df566b8-mgngk"] Feb 15 20:35:52 crc kubenswrapper[4735]: I0215 20:35:52.969489 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c7df566b8-mgngk" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api-log" containerID="cri-o://6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5" gracePeriod=30 Feb 15 20:35:52 crc kubenswrapper[4735]: I0215 20:35:52.970335 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c7df566b8-mgngk" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api" containerID="cri-o://5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e" gracePeriod=30 Feb 15 20:35:53 crc kubenswrapper[4735]: I0215 20:35:53.017364 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c7df566b8-mgngk" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": EOF" Feb 15 20:35:53 crc kubenswrapper[4735]: I0215 20:35:53.017576 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c7df566b8-mgngk" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": EOF" Feb 15 20:35:53 crc kubenswrapper[4735]: I0215 20:35:53.837639 4735 generic.go:334] "Generic (PLEG): container finished" podID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerID="6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5" exitCode=143 Feb 15 20:35:53 crc kubenswrapper[4735]: I0215 20:35:53.837717 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c7df566b8-mgngk" event={"ID":"c5391fc9-2f13-446e-9b98-8f453c23393f","Type":"ContainerDied","Data":"6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5"} Feb 15 20:35:54 crc kubenswrapper[4735]: I0215 20:35:54.777606 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:35:54 crc kubenswrapper[4735]: I0215 20:35:54.777886 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:35:54 crc kubenswrapper[4735]: I0215 20:35:54.778682 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"509c549dfe02bd37ca623752f3760afb9e5b4c41da36b238f99fb78d66aac256"} pod="openstack/horizon-5d466d479d-wdxj6" containerMessage="Container horizon failed startup probe, will be restarted" Feb 15 20:35:54 crc kubenswrapper[4735]: I0215 20:35:54.778718 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" containerID="cri-o://509c549dfe02bd37ca623752f3760afb9e5b4c41da36b238f99fb78d66aac256" gracePeriod=30 Feb 15 20:35:54 crc kubenswrapper[4735]: I0215 20:35:54.846856 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c647fbf5b-7zs9r" podUID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 15 20:35:54 crc kubenswrapper[4735]: I0215 20:35:54.846922 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:35:54 crc kubenswrapper[4735]: I0215 20:35:54.847954 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"063addf10589b9afd3c588e4171de6d504c19a3cbf44f266f86a25dfe5ca99e3"} pod="openstack/horizon-c647fbf5b-7zs9r" containerMessage="Container horizon failed startup probe, will be restarted" Feb 15 20:35:54 crc kubenswrapper[4735]: I0215 20:35:54.847992 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c647fbf5b-7zs9r" podUID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerName="horizon" containerID="cri-o://063addf10589b9afd3c588e4171de6d504c19a3cbf44f266f86a25dfe5ca99e3" gracePeriod=30 Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.179713 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6htdh" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.284615 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxvj5\" (UniqueName: \"kubernetes.io/projected/ec4fe5e3-ca92-4814-8921-62117a037281-kube-api-access-kxvj5\") pod \"ec4fe5e3-ca92-4814-8921-62117a037281\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.284742 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-combined-ca-bundle\") pod \"ec4fe5e3-ca92-4814-8921-62117a037281\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.284856 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-db-sync-config-data\") pod \"ec4fe5e3-ca92-4814-8921-62117a037281\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.284923 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-scripts\") pod \"ec4fe5e3-ca92-4814-8921-62117a037281\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.284981 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec4fe5e3-ca92-4814-8921-62117a037281-etc-machine-id\") pod \"ec4fe5e3-ca92-4814-8921-62117a037281\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.284999 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-config-data\") pod \"ec4fe5e3-ca92-4814-8921-62117a037281\" (UID: \"ec4fe5e3-ca92-4814-8921-62117a037281\") " Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.287498 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec4fe5e3-ca92-4814-8921-62117a037281-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ec4fe5e3-ca92-4814-8921-62117a037281" (UID: "ec4fe5e3-ca92-4814-8921-62117a037281"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.294932 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec4fe5e3-ca92-4814-8921-62117a037281-kube-api-access-kxvj5" (OuterVolumeSpecName: "kube-api-access-kxvj5") pod "ec4fe5e3-ca92-4814-8921-62117a037281" (UID: "ec4fe5e3-ca92-4814-8921-62117a037281"). InnerVolumeSpecName "kube-api-access-kxvj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.296471 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ec4fe5e3-ca92-4814-8921-62117a037281" (UID: "ec4fe5e3-ca92-4814-8921-62117a037281"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.312716 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-scripts" (OuterVolumeSpecName: "scripts") pod "ec4fe5e3-ca92-4814-8921-62117a037281" (UID: "ec4fe5e3-ca92-4814-8921-62117a037281"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.321770 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec4fe5e3-ca92-4814-8921-62117a037281" (UID: "ec4fe5e3-ca92-4814-8921-62117a037281"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.382388 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-config-data" (OuterVolumeSpecName: "config-data") pod "ec4fe5e3-ca92-4814-8921-62117a037281" (UID: "ec4fe5e3-ca92-4814-8921-62117a037281"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.387060 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.387083 4735 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.387092 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.387102 4735 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec4fe5e3-ca92-4814-8921-62117a037281-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.387110 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec4fe5e3-ca92-4814-8921-62117a037281-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.387118 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxvj5\" (UniqueName: \"kubernetes.io/projected/ec4fe5e3-ca92-4814-8921-62117a037281-kube-api-access-kxvj5\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.860132 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6htdh" event={"ID":"ec4fe5e3-ca92-4814-8921-62117a037281","Type":"ContainerDied","Data":"913e4bd190dbe2ba42c28044f10622cd9aa46d4ddf427984c1218b1e14c4468f"} Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.860170 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="913e4bd190dbe2ba42c28044f10622cd9aa46d4ddf427984c1218b1e14c4468f" Feb 15 20:35:55 crc kubenswrapper[4735]: I0215 20:35:55.860270 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6htdh" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.108599 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.109014 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-62lg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(5d7a00e9-2271-49b1-a74f-51f5f668995a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.110814 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.254197 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.281888 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.294199 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317108 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-ovndb-tls-certs\") pod \"2ef1315b-689a-4bab-9e16-0f767cd22170\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317163 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faf23e04-da47-4121-9e55-d003d8d6dcfe-logs\") pod \"faf23e04-da47-4121-9e55-d003d8d6dcfe\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317230 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-horizon-secret-key\") pod \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317270 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-httpd-config\") pod \"2ef1315b-689a-4bab-9e16-0f767cd22170\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317384 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-combined-ca-bundle\") pod \"2ef1315b-689a-4bab-9e16-0f767cd22170\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317412 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-scripts\") pod \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317469 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-config-data\") pod \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317490 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-internal-tls-certs\") pod \"2ef1315b-689a-4bab-9e16-0f767cd22170\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317515 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-config-data\") pod \"faf23e04-da47-4121-9e55-d003d8d6dcfe\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317543 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-config\") pod \"2ef1315b-689a-4bab-9e16-0f767cd22170\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317569 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxmf6\" (UniqueName: \"kubernetes.io/projected/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-kube-api-access-mxmf6\") pod \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317597 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-public-tls-certs\") pod \"2ef1315b-689a-4bab-9e16-0f767cd22170\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317638 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdksf\" (UniqueName: \"kubernetes.io/projected/2ef1315b-689a-4bab-9e16-0f767cd22170-kube-api-access-vdksf\") pod \"2ef1315b-689a-4bab-9e16-0f767cd22170\" (UID: \"2ef1315b-689a-4bab-9e16-0f767cd22170\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317664 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/faf23e04-da47-4121-9e55-d003d8d6dcfe-horizon-secret-key\") pod \"faf23e04-da47-4121-9e55-d003d8d6dcfe\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317702 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-logs\") pod \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\" (UID: \"a6f89bfa-6fb3-40d0-9741-cfddddb9a685\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317736 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfbp4\" (UniqueName: \"kubernetes.io/projected/faf23e04-da47-4121-9e55-d003d8d6dcfe-kube-api-access-mfbp4\") pod \"faf23e04-da47-4121-9e55-d003d8d6dcfe\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.317777 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-scripts\") pod \"faf23e04-da47-4121-9e55-d003d8d6dcfe\" (UID: \"faf23e04-da47-4121-9e55-d003d8d6dcfe\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.318573 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faf23e04-da47-4121-9e55-d003d8d6dcfe-logs" (OuterVolumeSpecName: "logs") pod "faf23e04-da47-4121-9e55-d003d8d6dcfe" (UID: "faf23e04-da47-4121-9e55-d003d8d6dcfe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.335425 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.336266 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-logs" (OuterVolumeSpecName: "logs") pod "a6f89bfa-6fb3-40d0-9741-cfddddb9a685" (UID: "a6f89bfa-6fb3-40d0-9741-cfddddb9a685"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.359193 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faf23e04-da47-4121-9e55-d003d8d6dcfe-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "faf23e04-da47-4121-9e55-d003d8d6dcfe" (UID: "faf23e04-da47-4121-9e55-d003d8d6dcfe"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.359199 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "2ef1315b-689a-4bab-9e16-0f767cd22170" (UID: "2ef1315b-689a-4bab-9e16-0f767cd22170"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.361572 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a6f89bfa-6fb3-40d0-9741-cfddddb9a685" (UID: "a6f89bfa-6fb3-40d0-9741-cfddddb9a685"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.361674 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ef1315b-689a-4bab-9e16-0f767cd22170-kube-api-access-vdksf" (OuterVolumeSpecName: "kube-api-access-vdksf") pod "2ef1315b-689a-4bab-9e16-0f767cd22170" (UID: "2ef1315b-689a-4bab-9e16-0f767cd22170"). InnerVolumeSpecName "kube-api-access-vdksf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.363544 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-kube-api-access-mxmf6" (OuterVolumeSpecName: "kube-api-access-mxmf6") pod "a6f89bfa-6fb3-40d0-9741-cfddddb9a685" (UID: "a6f89bfa-6fb3-40d0-9741-cfddddb9a685"). InnerVolumeSpecName "kube-api-access-mxmf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.379447 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faf23e04-da47-4121-9e55-d003d8d6dcfe-kube-api-access-mfbp4" (OuterVolumeSpecName: "kube-api-access-mfbp4") pod "faf23e04-da47-4121-9e55-d003d8d6dcfe" (UID: "faf23e04-da47-4121-9e55-d003d8d6dcfe"). InnerVolumeSpecName "kube-api-access-mfbp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.419616 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1be12817-7c08-4d66-938d-98ce9f79e4fd-horizon-secret-key\") pod \"1be12817-7c08-4d66-938d-98ce9f79e4fd\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.419667 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d787c\" (UniqueName: \"kubernetes.io/projected/1be12817-7c08-4d66-938d-98ce9f79e4fd-kube-api-access-d787c\") pod \"1be12817-7c08-4d66-938d-98ce9f79e4fd\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.419730 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1be12817-7c08-4d66-938d-98ce9f79e4fd-logs\") pod \"1be12817-7c08-4d66-938d-98ce9f79e4fd\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420036 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1be12817-7c08-4d66-938d-98ce9f79e4fd-logs" (OuterVolumeSpecName: "logs") pod "1be12817-7c08-4d66-938d-98ce9f79e4fd" (UID: "1be12817-7c08-4d66-938d-98ce9f79e4fd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420076 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-scripts\") pod \"1be12817-7c08-4d66-938d-98ce9f79e4fd\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420132 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-config-data\") pod \"1be12817-7c08-4d66-938d-98ce9f79e4fd\" (UID: \"1be12817-7c08-4d66-938d-98ce9f79e4fd\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420607 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdksf\" (UniqueName: \"kubernetes.io/projected/2ef1315b-689a-4bab-9e16-0f767cd22170-kube-api-access-vdksf\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420623 4735 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/faf23e04-da47-4121-9e55-d003d8d6dcfe-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420632 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420641 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfbp4\" (UniqueName: \"kubernetes.io/projected/faf23e04-da47-4121-9e55-d003d8d6dcfe-kube-api-access-mfbp4\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420651 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faf23e04-da47-4121-9e55-d003d8d6dcfe-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420660 4735 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420667 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.420676 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxmf6\" (UniqueName: \"kubernetes.io/projected/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-kube-api-access-mxmf6\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.462154 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-scripts" (OuterVolumeSpecName: "scripts") pod "faf23e04-da47-4121-9e55-d003d8d6dcfe" (UID: "faf23e04-da47-4121-9e55-d003d8d6dcfe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.473057 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1be12817-7c08-4d66-938d-98ce9f79e4fd-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1be12817-7c08-4d66-938d-98ce9f79e4fd" (UID: "1be12817-7c08-4d66-938d-98ce9f79e4fd"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.477892 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1be12817-7c08-4d66-938d-98ce9f79e4fd-kube-api-access-d787c" (OuterVolumeSpecName: "kube-api-access-d787c") pod "1be12817-7c08-4d66-938d-98ce9f79e4fd" (UID: "1be12817-7c08-4d66-938d-98ce9f79e4fd"). InnerVolumeSpecName "kube-api-access-d787c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.508249 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.528662 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.528700 4735 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1be12817-7c08-4d66-938d-98ce9f79e4fd-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.528713 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d787c\" (UniqueName: \"kubernetes.io/projected/1be12817-7c08-4d66-938d-98ce9f79e4fd-kube-api-access-d787c\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.528725 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1be12817-7c08-4d66-938d-98ce9f79e4fd-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.575627 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576073 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerName="horizon" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576090 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerName="horizon" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576108 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerName="neutron-httpd" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576116 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerName="neutron-httpd" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576128 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerName="neutron-api" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576135 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerName="neutron-api" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576147 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerName="horizon-log" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576152 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerName="horizon-log" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576165 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerName="dnsmasq-dns" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576170 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerName="dnsmasq-dns" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576179 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec4fe5e3-ca92-4814-8921-62117a037281" containerName="cinder-db-sync" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576185 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec4fe5e3-ca92-4814-8921-62117a037281" containerName="cinder-db-sync" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576201 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerName="init" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576207 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerName="init" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576219 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerName="horizon" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576225 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerName="horizon" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576241 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerName="horizon-log" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576246 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerName="horizon-log" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576258 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerName="horizon" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576264 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerName="horizon" Feb 15 20:35:56 crc kubenswrapper[4735]: E0215 20:35:56.576272 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerName="horizon-log" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576277 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerName="horizon-log" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576433 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerName="horizon-log" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576449 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerName="horizon" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576461 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" containerName="horizon" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576470 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerName="neutron-api" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576481 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerName="horizon" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576490 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec4fe5e3-ca92-4814-8921-62117a037281" containerName="cinder-db-sync" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576499 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed582df2-f4b9-4c7d-9f67-695b388d649d" containerName="dnsmasq-dns" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576530 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="faf23e04-da47-4121-9e55-d003d8d6dcfe" containerName="horizon-log" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576537 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" containerName="neutron-httpd" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.576547 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1be12817-7c08-4d66-938d-98ce9f79e4fd" containerName="horizon-log" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.577463 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.604748 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.604997 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.605105 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-76wr6" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.605206 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.608095 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-config-data" (OuterVolumeSpecName: "config-data") pod "a6f89bfa-6fb3-40d0-9741-cfddddb9a685" (UID: "a6f89bfa-6fb3-40d0-9741-cfddddb9a685"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.620929 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-config-data" (OuterVolumeSpecName: "config-data") pod "faf23e04-da47-4121-9e55-d003d8d6dcfe" (UID: "faf23e04-da47-4121-9e55-d003d8d6dcfe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.635909 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-svc\") pod \"ed582df2-f4b9-4c7d-9f67-695b388d649d\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.635964 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-config\") pod \"ed582df2-f4b9-4c7d-9f67-695b388d649d\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636091 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-swift-storage-0\") pod \"ed582df2-f4b9-4c7d-9f67-695b388d649d\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636164 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-sb\") pod \"ed582df2-f4b9-4c7d-9f67-695b388d649d\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636187 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-nb\") pod \"ed582df2-f4b9-4c7d-9f67-695b388d649d\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636226 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plmt6\" (UniqueName: \"kubernetes.io/projected/ed582df2-f4b9-4c7d-9f67-695b388d649d-kube-api-access-plmt6\") pod \"ed582df2-f4b9-4c7d-9f67-695b388d649d\" (UID: \"ed582df2-f4b9-4c7d-9f67-695b388d649d\") " Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636765 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636793 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-scripts\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636919 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636957 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7wl2\" (UniqueName: \"kubernetes.io/projected/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-kube-api-access-c7wl2\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.636988 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.637028 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.637039 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/faf23e04-da47-4121-9e55-d003d8d6dcfe-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.659699 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.660045 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-scripts" (OuterVolumeSpecName: "scripts") pod "1be12817-7c08-4d66-938d-98ce9f79e4fd" (UID: "1be12817-7c08-4d66-938d-98ce9f79e4fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.681494 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-scripts" (OuterVolumeSpecName: "scripts") pod "a6f89bfa-6fb3-40d0-9741-cfddddb9a685" (UID: "a6f89bfa-6fb3-40d0-9741-cfddddb9a685"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.739038 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7wl2\" (UniqueName: \"kubernetes.io/projected/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-kube-api-access-c7wl2\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.739266 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.739428 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.739516 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-scripts\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.739676 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.739782 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.739902 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f89bfa-6fb3-40d0-9741-cfddddb9a685-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.740027 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.742785 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-config-data" (OuterVolumeSpecName: "config-data") pod "1be12817-7c08-4d66-938d-98ce9f79e4fd" (UID: "1be12817-7c08-4d66-938d-98ce9f79e4fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.750821 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.753085 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.755419 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-scripts\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.755642 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.756919 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ef1315b-689a-4bab-9e16-0f767cd22170" (UID: "2ef1315b-689a-4bab-9e16-0f767cd22170"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.756961 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed582df2-f4b9-4c7d-9f67-695b388d649d-kube-api-access-plmt6" (OuterVolumeSpecName: "kube-api-access-plmt6") pod "ed582df2-f4b9-4c7d-9f67-695b388d649d" (UID: "ed582df2-f4b9-4c7d-9f67-695b388d649d"). InnerVolumeSpecName "kube-api-access-plmt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.759608 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.774375 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2ef1315b-689a-4bab-9e16-0f767cd22170" (UID: "2ef1315b-689a-4bab-9e16-0f767cd22170"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.776719 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-sjz6k"] Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.787596 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.791681 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-config" (OuterVolumeSpecName: "config") pod "2ef1315b-689a-4bab-9e16-0f767cd22170" (UID: "2ef1315b-689a-4bab-9e16-0f767cd22170"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.792009 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2ef1315b-689a-4bab-9e16-0f767cd22170" (UID: "2ef1315b-689a-4bab-9e16-0f767cd22170"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.795036 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-sjz6k"] Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.834236 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7wl2\" (UniqueName: \"kubernetes.io/projected/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-kube-api-access-c7wl2\") pod \"cinder-scheduler-0\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " pod="openstack/cinder-scheduler-0" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872467 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872554 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872612 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjxf9\" (UniqueName: \"kubernetes.io/projected/caed4f3d-6a00-40e6-aa12-dca360010f40-kube-api-access-fjxf9\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872726 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-svc\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872749 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-config\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872827 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1be12817-7c08-4d66-938d-98ce9f79e4fd-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872839 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872848 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872856 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872863 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.872873 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plmt6\" (UniqueName: \"kubernetes.io/projected/ed582df2-f4b9-4c7d-9f67-695b388d649d-kube-api-access-plmt6\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:56 crc kubenswrapper[4735]: I0215 20:35:56.992968 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed582df2-f4b9-4c7d-9f67-695b388d649d" (UID: "ed582df2-f4b9-4c7d-9f67-695b388d649d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.010234 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.011346 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6898675dc5-fh44t" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.076566 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557bd9599f-5lqht" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.108844 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "2ef1315b-689a-4bab-9e16-0f767cd22170" (UID: "2ef1315b-689a-4bab-9e16-0f767cd22170"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.111198 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.111290 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjxf9\" (UniqueName: \"kubernetes.io/projected/caed4f3d-6a00-40e6-aa12-dca360010f40-kube-api-access-fjxf9\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.113545 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.114058 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-svc\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.114097 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-config\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.114233 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.114405 4735 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ef1315b-689a-4bab-9e16-0f767cd22170-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.114424 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.114983 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.115513 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-svc\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.112042 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.129684 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.130351 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-config\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.131754 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.169599 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-config" (OuterVolumeSpecName: "config") pod "ed582df2-f4b9-4c7d-9f67-695b388d649d" (UID: "ed582df2-f4b9-4c7d-9f67-695b388d649d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.179460 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjxf9\" (UniqueName: \"kubernetes.io/projected/caed4f3d-6a00-40e6-aa12-dca360010f40-kube-api-access-fjxf9\") pod \"dnsmasq-dns-6578955fd5-sjz6k\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.180533 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6898675dc5-fh44t" event={"ID":"a6f89bfa-6fb3-40d0-9741-cfddddb9a685","Type":"ContainerDied","Data":"cacdff0549f62ad06f144926de14a801428227fe5e9b143386038fe067d4ba90"} Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.180588 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.183175 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.183201 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557bd9599f-5lqht" event={"ID":"2ef1315b-689a-4bab-9e16-0f767cd22170","Type":"ContainerDied","Data":"5d8c20b9735ab059862611d7ddf724fa4ba532cbd97efd1200d69b4ff9693179"} Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.183216 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-8ggxh" event={"ID":"ed582df2-f4b9-4c7d-9f67-695b388d649d","Type":"ContainerDied","Data":"95deee861412ac7aaf695eda38712435721b54c678d11f79608b7f2f85e4ffb2"} Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.183311 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.200758 4735 scope.go:117] "RemoveContainer" containerID="a3e86739b52ad23dbac1a7f4cfcf2a070a9b0f25ee1f6dad3b0794fcaa858d61" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.203758 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.214526 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ed582df2-f4b9-4c7d-9f67-695b388d649d" (UID: "ed582df2-f4b9-4c7d-9f67-695b388d649d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.216180 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f7d59545-hbprm" event={"ID":"1be12817-7c08-4d66-938d-98ce9f79e4fd","Type":"ContainerDied","Data":"6a47f18243090f01153e07f9b9083ee47fa55dc44e5b50014bdc718407908cc8"} Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.216295 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74f7d59545-hbprm" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.239830 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k7gb\" (UniqueName: \"kubernetes.io/projected/107aa5db-f0d0-4817-8d2d-12da89492473-kube-api-access-9k7gb\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.239902 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/107aa5db-f0d0-4817-8d2d-12da89492473-etc-machine-id\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.239961 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.242404 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data-custom\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.251539 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed582df2-f4b9-4c7d-9f67-695b388d649d" (UID: "ed582df2-f4b9-4c7d-9f67-695b388d649d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.257319 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-scripts\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.268496 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.274250 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fdf549c7-xpft7"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.275023 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/107aa5db-f0d0-4817-8d2d-12da89492473-logs\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.275409 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.275422 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.275527 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.286960 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74dd7854c7-t2t6r" event={"ID":"faf23e04-da47-4121-9e55-d003d8d6dcfe","Type":"ContainerDied","Data":"d861c6c2fb63c9717fc81d6f9f73ebc1c1006699b23adde1d101b4d0bb28ba8f"} Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.287022 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerName="ceilometer-notification-agent" containerID="cri-o://f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99" gracePeriod=30 Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.287076 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerName="sg-core" containerID="cri-o://5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804" gracePeriod=30 Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.287170 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74dd7854c7-t2t6r" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.294184 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed582df2-f4b9-4c7d-9f67-695b388d649d" (UID: "ed582df2-f4b9-4c7d-9f67-695b388d649d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.308170 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-557bd9599f-5lqht"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.318786 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-557bd9599f-5lqht"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.376793 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.376843 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data-custom\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.376879 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-scripts\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.376912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.376938 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/107aa5db-f0d0-4817-8d2d-12da89492473-logs\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.376981 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k7gb\" (UniqueName: \"kubernetes.io/projected/107aa5db-f0d0-4817-8d2d-12da89492473-kube-api-access-9k7gb\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.377024 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/107aa5db-f0d0-4817-8d2d-12da89492473-etc-machine-id\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.377082 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed582df2-f4b9-4c7d-9f67-695b388d649d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.377130 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/107aa5db-f0d0-4817-8d2d-12da89492473-etc-machine-id\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.380489 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.387623 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data-custom\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.387737 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/107aa5db-f0d0-4817-8d2d-12da89492473-logs\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.387838 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-scripts\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.390115 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.421353 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k7gb\" (UniqueName: \"kubernetes.io/projected/107aa5db-f0d0-4817-8d2d-12da89492473-kube-api-access-9k7gb\") pod \"cinder-api-0\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.422376 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.505386 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c7df566b8-mgngk" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:54876->10.217.0.166:9311: read: connection reset by peer" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.505405 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c7df566b8-mgngk" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:54892->10.217.0.166:9311: read: connection reset by peer" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.521155 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c7df566b8-mgngk" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": dial tcp 10.217.0.166:9311: connect: connection refused" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.521241 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.607770 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.676444 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74f7d59545-hbprm"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.685253 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-74f7d59545-hbprm"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.696815 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74dd7854c7-t2t6r"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.737038 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-74dd7854c7-t2t6r"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.751938 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6898675dc5-fh44t"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.780340 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6898675dc5-fh44t"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.790021 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8ggxh"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.797932 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8ggxh"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.860322 4735 scope.go:117] "RemoveContainer" containerID="5c3c8f494e812ff4a8c0aee1120fe0bd32c14ec07a9effb34a54ef3370575760" Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.904583 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 15 20:35:57 crc kubenswrapper[4735]: I0215 20:35:57.906640 4735 scope.go:117] "RemoveContainer" containerID="eb9e7b1d420557821175350c51326aa947abdf4c0c45729785ed5d87a383f30c" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.184080 4735 scope.go:117] "RemoveContainer" containerID="7a7b24352b641cb77fad01a59540c8459a7c1234dcf8766c7319d7e46daa5006" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.315501 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-sjz6k"] Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.315576 4735 scope.go:117] "RemoveContainer" containerID="dfe66441eaccfa6b08bdac1ca3e8b4ad52dc69c9543ac02fe7622c5f3c0ce6b0" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.317230 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.323853 4735 generic.go:334] "Generic (PLEG): container finished" podID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerID="5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e" exitCode=0 Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.323959 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c7df566b8-mgngk" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.324085 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c7df566b8-mgngk" event={"ID":"c5391fc9-2f13-446e-9b98-8f453c23393f","Type":"ContainerDied","Data":"5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e"} Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.324123 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c7df566b8-mgngk" event={"ID":"c5391fc9-2f13-446e-9b98-8f453c23393f","Type":"ContainerDied","Data":"13cd91783b5dd561d1cd0c360c9e3a96028e26b686a692190518164fa8b6d4e1"} Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.329201 4735 generic.go:334] "Generic (PLEG): container finished" podID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerID="5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804" exitCode=2 Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.329574 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d7a00e9-2271-49b1-a74f-51f5f668995a","Type":"ContainerDied","Data":"5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804"} Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.338912 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fdf549c7-xpft7" event={"ID":"33cc02bd-6580-4d99-9c26-5f0c222e6461","Type":"ContainerStarted","Data":"6471931dbdeda23c535740d420e84af135ecfc7ddf696a6a284f494a3b9ca34f"} Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.338967 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fdf549c7-xpft7" event={"ID":"33cc02bd-6580-4d99-9c26-5f0c222e6461","Type":"ContainerStarted","Data":"6373a729e45d77762f52a5eab084a78174637241ec784c32da25671f88bcd20f"} Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.349744 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992","Type":"ContainerStarted","Data":"3e7a0514ad4979346d958611ee463ca2f852daea2529fcc93844bfc9d2cd8702"} Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.373663 4735 scope.go:117] "RemoveContainer" containerID="82c48854bc164c771936514231eff3628e6b2bc61416220bc8cee0f4bab04980" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.408115 4735 scope.go:117] "RemoveContainer" containerID="10a22afe76edb65bfb324a00acf60c68c9864393f0fe03999f09757222e33183" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.429790 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lscs\" (UniqueName: \"kubernetes.io/projected/c5391fc9-2f13-446e-9b98-8f453c23393f-kube-api-access-8lscs\") pod \"c5391fc9-2f13-446e-9b98-8f453c23393f\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.429840 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data\") pod \"c5391fc9-2f13-446e-9b98-8f453c23393f\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.429897 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-combined-ca-bundle\") pod \"c5391fc9-2f13-446e-9b98-8f453c23393f\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.429916 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5391fc9-2f13-446e-9b98-8f453c23393f-logs\") pod \"c5391fc9-2f13-446e-9b98-8f453c23393f\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.429965 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data-custom\") pod \"c5391fc9-2f13-446e-9b98-8f453c23393f\" (UID: \"c5391fc9-2f13-446e-9b98-8f453c23393f\") " Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.430617 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5391fc9-2f13-446e-9b98-8f453c23393f-logs" (OuterVolumeSpecName: "logs") pod "c5391fc9-2f13-446e-9b98-8f453c23393f" (UID: "c5391fc9-2f13-446e-9b98-8f453c23393f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.441479 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c5391fc9-2f13-446e-9b98-8f453c23393f" (UID: "c5391fc9-2f13-446e-9b98-8f453c23393f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.445302 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5391fc9-2f13-446e-9b98-8f453c23393f-kube-api-access-8lscs" (OuterVolumeSpecName: "kube-api-access-8lscs") pod "c5391fc9-2f13-446e-9b98-8f453c23393f" (UID: "c5391fc9-2f13-446e-9b98-8f453c23393f"). InnerVolumeSpecName "kube-api-access-8lscs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.488262 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5391fc9-2f13-446e-9b98-8f453c23393f" (UID: "c5391fc9-2f13-446e-9b98-8f453c23393f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.531616 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lscs\" (UniqueName: \"kubernetes.io/projected/c5391fc9-2f13-446e-9b98-8f453c23393f-kube-api-access-8lscs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.531647 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.531657 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5391fc9-2f13-446e-9b98-8f453c23393f-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.531667 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.537072 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data" (OuterVolumeSpecName: "config-data") pod "c5391fc9-2f13-446e-9b98-8f453c23393f" (UID: "c5391fc9-2f13-446e-9b98-8f453c23393f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.603990 4735 scope.go:117] "RemoveContainer" containerID="81739c66d31bfab9b2a5e97080b0eccbabd62978dd60f71980b7bfc6c8bc8839" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.611334 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.632743 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5391fc9-2f13-446e-9b98-8f453c23393f-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.676088 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c7df566b8-mgngk"] Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.693011 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7c7df566b8-mgngk"] Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.822246 4735 scope.go:117] "RemoveContainer" containerID="363de19826e86b82ff64c234ca3ec5e5b9bfa1a8265f4b25fb7531f400e00c7c" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.965378 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1be12817-7c08-4d66-938d-98ce9f79e4fd" path="/var/lib/kubelet/pods/1be12817-7c08-4d66-938d-98ce9f79e4fd/volumes" Feb 15 20:35:58 crc kubenswrapper[4735]: I0215 20:35:58.988632 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ef1315b-689a-4bab-9e16-0f767cd22170" path="/var/lib/kubelet/pods/2ef1315b-689a-4bab-9e16-0f767cd22170/volumes" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.008353 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6f89bfa-6fb3-40d0-9741-cfddddb9a685" path="/var/lib/kubelet/pods/a6f89bfa-6fb3-40d0-9741-cfddddb9a685/volumes" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.009142 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" path="/var/lib/kubelet/pods/c5391fc9-2f13-446e-9b98-8f453c23393f/volumes" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.009822 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed582df2-f4b9-4c7d-9f67-695b388d649d" path="/var/lib/kubelet/pods/ed582df2-f4b9-4c7d-9f67-695b388d649d/volumes" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.011397 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faf23e04-da47-4121-9e55-d003d8d6dcfe" path="/var/lib/kubelet/pods/faf23e04-da47-4121-9e55-d003d8d6dcfe/volumes" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.085216 4735 scope.go:117] "RemoveContainer" containerID="f295535d4a74689debfd09b839f03d0ba86e654fdda24a7807dcd72de4f9eb23" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.120387 4735 scope.go:117] "RemoveContainer" containerID="5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.125074 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.256064 4735 scope.go:117] "RemoveContainer" containerID="6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.348762 4735 scope.go:117] "RemoveContainer" containerID="5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e" Feb 15 20:35:59 crc kubenswrapper[4735]: E0215 20:35:59.349850 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e\": container with ID starting with 5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e not found: ID does not exist" containerID="5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.350108 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e"} err="failed to get container status \"5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e\": rpc error: code = NotFound desc = could not find container \"5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e\": container with ID starting with 5a54d9ebee41a65150dcb25b4de78958f61a5363b280875e3a721b29a3f90e1e not found: ID does not exist" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.350132 4735 scope.go:117] "RemoveContainer" containerID="6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5" Feb 15 20:35:59 crc kubenswrapper[4735]: E0215 20:35:59.354605 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5\": container with ID starting with 6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5 not found: ID does not exist" containerID="6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.354650 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5"} err="failed to get container status \"6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5\": rpc error: code = NotFound desc = could not find container \"6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5\": container with ID starting with 6f9b136d412fc55eb57972d2262d8089db5436f6e00cfe5b707f3c876e7662c5 not found: ID does not exist" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.369302 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fdf549c7-xpft7" event={"ID":"33cc02bd-6580-4d99-9c26-5f0c222e6461","Type":"ContainerStarted","Data":"fce8cf789d92aa2bc9957ef2436e84ee8488b9f9e49df5057162bf26e80f1e84"} Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.370623 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.409559 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"107aa5db-f0d0-4817-8d2d-12da89492473","Type":"ContainerStarted","Data":"5872ffbc7b59e977da9d19ee421270cdc341d454bac69a3db14c9d0c7ce65559"} Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.417727 4735 generic.go:334] "Generic (PLEG): container finished" podID="caed4f3d-6a00-40e6-aa12-dca360010f40" containerID="9a4f845c67566578f0c82ae18b26b62798e2aae0e42305d5b9a1d3b3ce600938" exitCode=0 Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.417787 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" event={"ID":"caed4f3d-6a00-40e6-aa12-dca360010f40","Type":"ContainerDied","Data":"9a4f845c67566578f0c82ae18b26b62798e2aae0e42305d5b9a1d3b3ce600938"} Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.417807 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" event={"ID":"caed4f3d-6a00-40e6-aa12-dca360010f40","Type":"ContainerStarted","Data":"01d135c02fa967bf93f3e2241095b19ec01d69cb083acc8596d42a54b126ccbd"} Feb 15 20:35:59 crc kubenswrapper[4735]: I0215 20:35:59.467662 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-fdf549c7-xpft7" podStartSLOduration=11.467638999 podStartE2EDuration="11.467638999s" podCreationTimestamp="2026-02-15 20:35:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:35:59.396579034 +0000 UTC m=+1167.262594747" watchObservedRunningTime="2026-02-15 20:35:59.467638999 +0000 UTC m=+1167.333654612" Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.455975 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992","Type":"ContainerStarted","Data":"155cbd5c1a2b149ea2cfdbb545a57313f9df4e96f5035271ba7286e61de55c67"} Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.459219 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"107aa5db-f0d0-4817-8d2d-12da89492473","Type":"ContainerStarted","Data":"be1dc831e7060f2c2a416d0785c4bfb76e59d77ebfaab5dd8be8cff4c319dd71"} Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.459258 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"107aa5db-f0d0-4817-8d2d-12da89492473","Type":"ContainerStarted","Data":"752d9da1cf37173f495550234a7f8e573937d3f4e6f9151ac82a2829ac2cd97f"} Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.459357 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="107aa5db-f0d0-4817-8d2d-12da89492473" containerName="cinder-api-log" containerID="cri-o://752d9da1cf37173f495550234a7f8e573937d3f4e6f9151ac82a2829ac2cd97f" gracePeriod=30 Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.459501 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="107aa5db-f0d0-4817-8d2d-12da89492473" containerName="cinder-api" containerID="cri-o://be1dc831e7060f2c2a416d0785c4bfb76e59d77ebfaab5dd8be8cff4c319dd71" gracePeriod=30 Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.459633 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.473910 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" event={"ID":"caed4f3d-6a00-40e6-aa12-dca360010f40","Type":"ContainerStarted","Data":"231fba9713abe30ab136dbf606840fd5142e31d9fa1838e0485a397e97043226"} Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.473966 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.484005 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.483986632 podStartE2EDuration="4.483986632s" podCreationTimestamp="2026-02-15 20:35:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:00.477823875 +0000 UTC m=+1168.343839498" watchObservedRunningTime="2026-02-15 20:36:00.483986632 +0000 UTC m=+1168.350002255" Feb 15 20:36:00 crc kubenswrapper[4735]: I0215 20:36:00.507240 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" podStartSLOduration=4.507220622 podStartE2EDuration="4.507220622s" podCreationTimestamp="2026-02-15 20:35:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:00.500423757 +0000 UTC m=+1168.366439380" watchObservedRunningTime="2026-02-15 20:36:00.507220622 +0000 UTC m=+1168.373236245" Feb 15 20:36:01 crc kubenswrapper[4735]: I0215 20:36:01.483466 4735 generic.go:334] "Generic (PLEG): container finished" podID="107aa5db-f0d0-4817-8d2d-12da89492473" containerID="752d9da1cf37173f495550234a7f8e573937d3f4e6f9151ac82a2829ac2cd97f" exitCode=143 Feb 15 20:36:01 crc kubenswrapper[4735]: I0215 20:36:01.483548 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"107aa5db-f0d0-4817-8d2d-12da89492473","Type":"ContainerDied","Data":"752d9da1cf37173f495550234a7f8e573937d3f4e6f9151ac82a2829ac2cd97f"} Feb 15 20:36:01 crc kubenswrapper[4735]: I0215 20:36:01.486901 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992","Type":"ContainerStarted","Data":"3125714211a6ad491c9cbe5aed193d273e3d63bd366826c59614d2d290f2ca6e"} Feb 15 20:36:01 crc kubenswrapper[4735]: I0215 20:36:01.514315 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.346742046 podStartE2EDuration="5.514294084s" podCreationTimestamp="2026-02-15 20:35:56 +0000 UTC" firstStartedPulling="2026-02-15 20:35:57.973739304 +0000 UTC m=+1165.839754927" lastFinishedPulling="2026-02-15 20:35:59.141291342 +0000 UTC m=+1167.007306965" observedRunningTime="2026-02-15 20:36:01.511418056 +0000 UTC m=+1169.377433679" watchObservedRunningTime="2026-02-15 20:36:01.514294084 +0000 UTC m=+1169.380309707" Feb 15 20:36:01 crc kubenswrapper[4735]: I0215 20:36:01.924557 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.012054 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.104307 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-sg-core-conf-yaml\") pod \"5d7a00e9-2271-49b1-a74f-51f5f668995a\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.104417 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-log-httpd\") pod \"5d7a00e9-2271-49b1-a74f-51f5f668995a\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.104474 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-config-data\") pod \"5d7a00e9-2271-49b1-a74f-51f5f668995a\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.104517 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-run-httpd\") pod \"5d7a00e9-2271-49b1-a74f-51f5f668995a\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.104558 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62lg9\" (UniqueName: \"kubernetes.io/projected/5d7a00e9-2271-49b1-a74f-51f5f668995a-kube-api-access-62lg9\") pod \"5d7a00e9-2271-49b1-a74f-51f5f668995a\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.104584 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-scripts\") pod \"5d7a00e9-2271-49b1-a74f-51f5f668995a\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.104610 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-combined-ca-bundle\") pod \"5d7a00e9-2271-49b1-a74f-51f5f668995a\" (UID: \"5d7a00e9-2271-49b1-a74f-51f5f668995a\") " Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.105270 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5d7a00e9-2271-49b1-a74f-51f5f668995a" (UID: "5d7a00e9-2271-49b1-a74f-51f5f668995a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.105482 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5d7a00e9-2271-49b1-a74f-51f5f668995a" (UID: "5d7a00e9-2271-49b1-a74f-51f5f668995a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.110758 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d7a00e9-2271-49b1-a74f-51f5f668995a-kube-api-access-62lg9" (OuterVolumeSpecName: "kube-api-access-62lg9") pod "5d7a00e9-2271-49b1-a74f-51f5f668995a" (UID: "5d7a00e9-2271-49b1-a74f-51f5f668995a"). InnerVolumeSpecName "kube-api-access-62lg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.113206 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-scripts" (OuterVolumeSpecName: "scripts") pod "5d7a00e9-2271-49b1-a74f-51f5f668995a" (UID: "5d7a00e9-2271-49b1-a74f-51f5f668995a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.144622 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5d7a00e9-2271-49b1-a74f-51f5f668995a" (UID: "5d7a00e9-2271-49b1-a74f-51f5f668995a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.161723 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d7a00e9-2271-49b1-a74f-51f5f668995a" (UID: "5d7a00e9-2271-49b1-a74f-51f5f668995a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.179278 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-config-data" (OuterVolumeSpecName: "config-data") pod "5d7a00e9-2271-49b1-a74f-51f5f668995a" (UID: "5d7a00e9-2271-49b1-a74f-51f5f668995a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.207049 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.207085 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62lg9\" (UniqueName: \"kubernetes.io/projected/5d7a00e9-2271-49b1-a74f-51f5f668995a-kube-api-access-62lg9\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.207095 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.207102 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.207110 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.207118 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d7a00e9-2271-49b1-a74f-51f5f668995a-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.207126 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7a00e9-2271-49b1-a74f-51f5f668995a-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.496508 4735 generic.go:334] "Generic (PLEG): container finished" podID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerID="f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99" exitCode=0 Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.496788 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.496702 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d7a00e9-2271-49b1-a74f-51f5f668995a","Type":"ContainerDied","Data":"f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99"} Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.496887 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d7a00e9-2271-49b1-a74f-51f5f668995a","Type":"ContainerDied","Data":"abe86075d510854b77bbb3615d54faf42689095565497865f59422016bdcbb4e"} Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.496910 4735 scope.go:117] "RemoveContainer" containerID="5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.522752 4735 scope.go:117] "RemoveContainer" containerID="f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.591662 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.602151 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.633264 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:02 crc kubenswrapper[4735]: E0215 20:36:02.633669 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerName="sg-core" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.633714 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerName="sg-core" Feb 15 20:36:02 crc kubenswrapper[4735]: E0215 20:36:02.633725 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.633732 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api" Feb 15 20:36:02 crc kubenswrapper[4735]: E0215 20:36:02.633757 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerName="ceilometer-notification-agent" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.633764 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerName="ceilometer-notification-agent" Feb 15 20:36:02 crc kubenswrapper[4735]: E0215 20:36:02.633774 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api-log" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.633779 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api-log" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.634012 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerName="ceilometer-notification-agent" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.634038 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" containerName="sg-core" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.634054 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api-log" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.634066 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5391fc9-2f13-446e-9b98-8f453c23393f" containerName="barbican-api" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.635623 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.640996 4735 scope.go:117] "RemoveContainer" containerID="5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.641727 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.641811 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 15 20:36:02 crc kubenswrapper[4735]: E0215 20:36:02.642321 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804\": container with ID starting with 5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804 not found: ID does not exist" containerID="5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.642362 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804"} err="failed to get container status \"5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804\": rpc error: code = NotFound desc = could not find container \"5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804\": container with ID starting with 5ed8a57a5e08f4ba542354cbd37041b5ae9ea5c5cdcf9bb7695d8b81d7317804 not found: ID does not exist" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.642388 4735 scope.go:117] "RemoveContainer" containerID="f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.648905 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:02 crc kubenswrapper[4735]: E0215 20:36:02.661073 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99\": container with ID starting with f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99 not found: ID does not exist" containerID="f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.661119 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99"} err="failed to get container status \"f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99\": rpc error: code = NotFound desc = could not find container \"f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99\": container with ID starting with f118e742fc85973bc8eb280da7ea33e3c3c3f2da0a50c41b33370a2f486c6e99 not found: ID does not exist" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.820093 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.820147 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.820229 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-run-httpd\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.820257 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-config-data\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.820279 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-log-httpd\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.820339 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xml2t\" (UniqueName: \"kubernetes.io/projected/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-kube-api-access-xml2t\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.820485 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-scripts\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.896875 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d7a00e9-2271-49b1-a74f-51f5f668995a" path="/var/lib/kubelet/pods/5d7a00e9-2271-49b1-a74f-51f5f668995a/volumes" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.921560 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xml2t\" (UniqueName: \"kubernetes.io/projected/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-kube-api-access-xml2t\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.921851 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-scripts\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.922455 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.922488 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.922519 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-run-httpd\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.922543 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-config-data\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.922558 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-log-httpd\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.922935 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-log-httpd\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.923123 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-run-httpd\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.930865 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-scripts\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.931026 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.931451 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.931494 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-config-data\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.940711 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xml2t\" (UniqueName: \"kubernetes.io/projected/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-kube-api-access-xml2t\") pod \"ceilometer-0\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " pod="openstack/ceilometer-0" Feb 15 20:36:02 crc kubenswrapper[4735]: I0215 20:36:02.975300 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:03 crc kubenswrapper[4735]: I0215 20:36:03.515138 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:04 crc kubenswrapper[4735]: I0215 20:36:04.520001 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerStarted","Data":"8d114b6173c49bd9cb2de87ba812abf6cde86c1ef7d83bcbe0aaff2a24425584"} Feb 15 20:36:04 crc kubenswrapper[4735]: I0215 20:36:04.520767 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerStarted","Data":"d81b7261a3462028a97a77e2a7b77d8f13094f5d866f42e0cf538bd7412e1013"} Feb 15 20:36:05 crc kubenswrapper[4735]: I0215 20:36:05.528284 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerStarted","Data":"563cd8683e20b062dc2adde56b23ee47afd98fa9e99b53d61e87b491b24d6835"} Feb 15 20:36:06 crc kubenswrapper[4735]: I0215 20:36:06.537740 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerStarted","Data":"6c226ff20ed0038cef94ee50f1214e3a0e63d9739012d1195ec9cb4b02b520c1"} Feb 15 20:36:06 crc kubenswrapper[4735]: I0215 20:36:06.935616 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.232458 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.413364 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.434327 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.562214 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.571092 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerName="probe" containerID="cri-o://3125714211a6ad491c9cbe5aed193d273e3d63bd366826c59614d2d290f2ca6e" gracePeriod=30 Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.571246 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerName="cinder-scheduler" containerID="cri-o://155cbd5c1a2b149ea2cfdbb545a57313f9df4e96f5035271ba7286e61de55c67" gracePeriod=30 Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.580881 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.585325 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6c5c4f5b9b-sg99c" Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.619975 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zrzbw"] Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.620223 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" podUID="c49c8247-c8d6-4f83-9b3c-565cb6fefe51" containerName="dnsmasq-dns" containerID="cri-o://f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15" gracePeriod=10 Feb 15 20:36:07 crc kubenswrapper[4735]: I0215 20:36:07.703305 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-58678bdfb8-bgfbt"] Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.353652 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.546311 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-sb\") pod \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.546387 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-nb\") pod \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.546451 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-svc\") pod \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.546491 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpkl6\" (UniqueName: \"kubernetes.io/projected/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-kube-api-access-lpkl6\") pod \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.546530 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-config\") pod \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.546553 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-swift-storage-0\") pod \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\" (UID: \"c49c8247-c8d6-4f83-9b3c-565cb6fefe51\") " Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.579069 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-kube-api-access-lpkl6" (OuterVolumeSpecName: "kube-api-access-lpkl6") pod "c49c8247-c8d6-4f83-9b3c-565cb6fefe51" (UID: "c49c8247-c8d6-4f83-9b3c-565cb6fefe51"). InnerVolumeSpecName "kube-api-access-lpkl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.608200 4735 generic.go:334] "Generic (PLEG): container finished" podID="c49c8247-c8d6-4f83-9b3c-565cb6fefe51" containerID="f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15" exitCode=0 Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.608503 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" event={"ID":"c49c8247-c8d6-4f83-9b3c-565cb6fefe51","Type":"ContainerDied","Data":"f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15"} Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.608587 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" event={"ID":"c49c8247-c8d6-4f83-9b3c-565cb6fefe51","Type":"ContainerDied","Data":"f6718d23c3977119b2b7cd3a510abd92955a136713ea763478c6ed74a3b5def5"} Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.608661 4735 scope.go:117] "RemoveContainer" containerID="f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.608843 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-zrzbw" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.642046 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerStarted","Data":"73e5563131beea3e291af24d55b00d086bcd2089890020e6c943154f24febfce"} Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.642191 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-58678bdfb8-bgfbt" podUID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerName="placement-log" containerID="cri-o://84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f" gracePeriod=30 Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.642289 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.642326 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-58678bdfb8-bgfbt" podUID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerName="placement-api" containerID="cri-o://991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226" gracePeriod=30 Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.647866 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpkl6\" (UniqueName: \"kubernetes.io/projected/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-kube-api-access-lpkl6\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.701931 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c49c8247-c8d6-4f83-9b3c-565cb6fefe51" (UID: "c49c8247-c8d6-4f83-9b3c-565cb6fefe51"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.712528 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.954655014 podStartE2EDuration="6.712512221s" podCreationTimestamp="2026-02-15 20:36:02 +0000 UTC" firstStartedPulling="2026-02-15 20:36:03.534995908 +0000 UTC m=+1171.401011531" lastFinishedPulling="2026-02-15 20:36:07.292853115 +0000 UTC m=+1175.158868738" observedRunningTime="2026-02-15 20:36:08.708875692 +0000 UTC m=+1176.574891315" watchObservedRunningTime="2026-02-15 20:36:08.712512221 +0000 UTC m=+1176.578527844" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.715419 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c49c8247-c8d6-4f83-9b3c-565cb6fefe51" (UID: "c49c8247-c8d6-4f83-9b3c-565cb6fefe51"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.733345 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-config" (OuterVolumeSpecName: "config") pod "c49c8247-c8d6-4f83-9b3c-565cb6fefe51" (UID: "c49c8247-c8d6-4f83-9b3c-565cb6fefe51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.752453 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c49c8247-c8d6-4f83-9b3c-565cb6fefe51" (UID: "c49c8247-c8d6-4f83-9b3c-565cb6fefe51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.753684 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.753700 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.753708 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.753716 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.758414 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c49c8247-c8d6-4f83-9b3c-565cb6fefe51" (UID: "c49c8247-c8d6-4f83-9b3c-565cb6fefe51"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.823449 4735 scope.go:117] "RemoveContainer" containerID="0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.856454 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c49c8247-c8d6-4f83-9b3c-565cb6fefe51-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.879650 4735 scope.go:117] "RemoveContainer" containerID="f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15" Feb 15 20:36:08 crc kubenswrapper[4735]: E0215 20:36:08.880046 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15\": container with ID starting with f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15 not found: ID does not exist" containerID="f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.880076 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15"} err="failed to get container status \"f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15\": rpc error: code = NotFound desc = could not find container \"f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15\": container with ID starting with f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15 not found: ID does not exist" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.880102 4735 scope.go:117] "RemoveContainer" containerID="0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263" Feb 15 20:36:08 crc kubenswrapper[4735]: E0215 20:36:08.880380 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263\": container with ID starting with 0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263 not found: ID does not exist" containerID="0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.880404 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263"} err="failed to get container status \"0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263\": rpc error: code = NotFound desc = could not find container \"0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263\": container with ID starting with 0f262cccd0f84d4c059856146cbc0ddfcbac673e50f03a8a8b6f8247d22ab263 not found: ID does not exist" Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.941557 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zrzbw"] Feb 15 20:36:08 crc kubenswrapper[4735]: I0215 20:36:08.949396 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zrzbw"] Feb 15 20:36:09 crc kubenswrapper[4735]: I0215 20:36:09.285901 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7cb4db7f5b-mrtp9" Feb 15 20:36:09 crc kubenswrapper[4735]: I0215 20:36:09.653956 4735 generic.go:334] "Generic (PLEG): container finished" podID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerID="3125714211a6ad491c9cbe5aed193d273e3d63bd366826c59614d2d290f2ca6e" exitCode=0 Feb 15 20:36:09 crc kubenswrapper[4735]: I0215 20:36:09.654065 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992","Type":"ContainerDied","Data":"3125714211a6ad491c9cbe5aed193d273e3d63bd366826c59614d2d290f2ca6e"} Feb 15 20:36:09 crc kubenswrapper[4735]: I0215 20:36:09.657208 4735 generic.go:334] "Generic (PLEG): container finished" podID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerID="84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f" exitCode=143 Feb 15 20:36:09 crc kubenswrapper[4735]: I0215 20:36:09.657261 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58678bdfb8-bgfbt" event={"ID":"e4ec4117-00bf-4584-a22e-f760637b52ba","Type":"ContainerDied","Data":"84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f"} Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.786985 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 15 20:36:10 crc kubenswrapper[4735]: E0215 20:36:10.787572 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c49c8247-c8d6-4f83-9b3c-565cb6fefe51" containerName="init" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.787588 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c49c8247-c8d6-4f83-9b3c-565cb6fefe51" containerName="init" Feb 15 20:36:10 crc kubenswrapper[4735]: E0215 20:36:10.787615 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c49c8247-c8d6-4f83-9b3c-565cb6fefe51" containerName="dnsmasq-dns" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.787621 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c49c8247-c8d6-4f83-9b3c-565cb6fefe51" containerName="dnsmasq-dns" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.787849 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c49c8247-c8d6-4f83-9b3c-565cb6fefe51" containerName="dnsmasq-dns" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.788467 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.795883 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.796054 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.796207 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-z84gm" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.796842 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.890768 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.890834 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.890923 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config-secret\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.890980 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qvpt\" (UniqueName: \"kubernetes.io/projected/d1cff902-6918-4f13-b1c2-cd5adad6650f-kube-api-access-5qvpt\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.897787 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c49c8247-c8d6-4f83-9b3c-565cb6fefe51" path="/var/lib/kubelet/pods/c49c8247-c8d6-4f83-9b3c-565cb6fefe51/volumes" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.992508 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.992904 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.993193 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config-secret\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.993312 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qvpt\" (UniqueName: \"kubernetes.io/projected/d1cff902-6918-4f13-b1c2-cd5adad6650f-kube-api-access-5qvpt\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.993907 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:10 crc kubenswrapper[4735]: I0215 20:36:10.994065 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:10.999932 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.000446 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config-secret\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.016539 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qvpt\" (UniqueName: \"kubernetes.io/projected/d1cff902-6918-4f13-b1c2-cd5adad6650f-kube-api-access-5qvpt\") pod \"openstackclient\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.108484 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.154034 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.182025 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.228332 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.229501 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.255587 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 15 20:36:11 crc kubenswrapper[4735]: E0215 20:36:11.291330 4735 log.go:32] "RunPodSandbox from runtime service failed" err=< Feb 15 20:36:11 crc kubenswrapper[4735]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_d1cff902-6918-4f13-b1c2-cd5adad6650f_0(10f6134a18a6f43efe69bba0ff33a0181dcc61187aa7a4baeac4f7b8c420330d): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"10f6134a18a6f43efe69bba0ff33a0181dcc61187aa7a4baeac4f7b8c420330d" Netns:"/var/run/netns/40b7866a-a9ca-43a1-8c7e-b23fda201574" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=10f6134a18a6f43efe69bba0ff33a0181dcc61187aa7a4baeac4f7b8c420330d;K8S_POD_UID=d1cff902-6918-4f13-b1c2-cd5adad6650f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/d1cff902-6918-4f13-b1c2-cd5adad6650f]: expected pod UID "d1cff902-6918-4f13-b1c2-cd5adad6650f" but got "0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15" from Kube API Feb 15 20:36:11 crc kubenswrapper[4735]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 15 20:36:11 crc kubenswrapper[4735]: > Feb 15 20:36:11 crc kubenswrapper[4735]: E0215 20:36:11.291557 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Feb 15 20:36:11 crc kubenswrapper[4735]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_d1cff902-6918-4f13-b1c2-cd5adad6650f_0(10f6134a18a6f43efe69bba0ff33a0181dcc61187aa7a4baeac4f7b8c420330d): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"10f6134a18a6f43efe69bba0ff33a0181dcc61187aa7a4baeac4f7b8c420330d" Netns:"/var/run/netns/40b7866a-a9ca-43a1-8c7e-b23fda201574" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=10f6134a18a6f43efe69bba0ff33a0181dcc61187aa7a4baeac4f7b8c420330d;K8S_POD_UID=d1cff902-6918-4f13-b1c2-cd5adad6650f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/d1cff902-6918-4f13-b1c2-cd5adad6650f]: expected pod UID "d1cff902-6918-4f13-b1c2-cd5adad6650f" but got "0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15" from Kube API Feb 15 20:36:11 crc kubenswrapper[4735]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 15 20:36:11 crc kubenswrapper[4735]: > pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.307836 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87lzd\" (UniqueName: \"kubernetes.io/projected/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-kube-api-access-87lzd\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.308144 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.308307 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-openstack-config-secret\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.308391 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-openstack-config\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.409552 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87lzd\" (UniqueName: \"kubernetes.io/projected/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-kube-api-access-87lzd\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.409595 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.409680 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-openstack-config-secret\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.409705 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-openstack-config\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.410776 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-openstack-config\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.415040 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.416728 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-openstack-config-secret\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.430482 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87lzd\" (UniqueName: \"kubernetes.io/projected/0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15-kube-api-access-87lzd\") pod \"openstackclient\" (UID: \"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15\") " pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.558386 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.749056 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.758264 4735 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d1cff902-6918-4f13-b1c2-cd5adad6650f" podUID="0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.792053 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.820304 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-combined-ca-bundle\") pod \"d1cff902-6918-4f13-b1c2-cd5adad6650f\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.820363 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config\") pod \"d1cff902-6918-4f13-b1c2-cd5adad6650f\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.820404 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qvpt\" (UniqueName: \"kubernetes.io/projected/d1cff902-6918-4f13-b1c2-cd5adad6650f-kube-api-access-5qvpt\") pod \"d1cff902-6918-4f13-b1c2-cd5adad6650f\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.820430 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config-secret\") pod \"d1cff902-6918-4f13-b1c2-cd5adad6650f\" (UID: \"d1cff902-6918-4f13-b1c2-cd5adad6650f\") " Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.821924 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d1cff902-6918-4f13-b1c2-cd5adad6650f" (UID: "d1cff902-6918-4f13-b1c2-cd5adad6650f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.825071 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d1cff902-6918-4f13-b1c2-cd5adad6650f" (UID: "d1cff902-6918-4f13-b1c2-cd5adad6650f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.831154 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1cff902-6918-4f13-b1c2-cd5adad6650f" (UID: "d1cff902-6918-4f13-b1c2-cd5adad6650f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.831474 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1cff902-6918-4f13-b1c2-cd5adad6650f-kube-api-access-5qvpt" (OuterVolumeSpecName: "kube-api-access-5qvpt") pod "d1cff902-6918-4f13-b1c2-cd5adad6650f" (UID: "d1cff902-6918-4f13-b1c2-cd5adad6650f"). InnerVolumeSpecName "kube-api-access-5qvpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.922176 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.922200 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.922209 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qvpt\" (UniqueName: \"kubernetes.io/projected/d1cff902-6918-4f13-b1c2-cd5adad6650f-kube-api-access-5qvpt\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:11 crc kubenswrapper[4735]: I0215 20:36:11.922219 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d1cff902-6918-4f13-b1c2-cd5adad6650f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.260940 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.424866 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.431548 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmtxk\" (UniqueName: \"kubernetes.io/projected/e4ec4117-00bf-4584-a22e-f760637b52ba-kube-api-access-pmtxk\") pod \"e4ec4117-00bf-4584-a22e-f760637b52ba\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.431598 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ec4117-00bf-4584-a22e-f760637b52ba-logs\") pod \"e4ec4117-00bf-4584-a22e-f760637b52ba\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.431643 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-config-data\") pod \"e4ec4117-00bf-4584-a22e-f760637b52ba\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.431688 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-internal-tls-certs\") pod \"e4ec4117-00bf-4584-a22e-f760637b52ba\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.431721 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-combined-ca-bundle\") pod \"e4ec4117-00bf-4584-a22e-f760637b52ba\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.431803 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-scripts\") pod \"e4ec4117-00bf-4584-a22e-f760637b52ba\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.431845 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-public-tls-certs\") pod \"e4ec4117-00bf-4584-a22e-f760637b52ba\" (UID: \"e4ec4117-00bf-4584-a22e-f760637b52ba\") " Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.432985 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4ec4117-00bf-4584-a22e-f760637b52ba-logs" (OuterVolumeSpecName: "logs") pod "e4ec4117-00bf-4584-a22e-f760637b52ba" (UID: "e4ec4117-00bf-4584-a22e-f760637b52ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.440004 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-scripts" (OuterVolumeSpecName: "scripts") pod "e4ec4117-00bf-4584-a22e-f760637b52ba" (UID: "e4ec4117-00bf-4584-a22e-f760637b52ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.457448 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ec4117-00bf-4584-a22e-f760637b52ba-kube-api-access-pmtxk" (OuterVolumeSpecName: "kube-api-access-pmtxk") pod "e4ec4117-00bf-4584-a22e-f760637b52ba" (UID: "e4ec4117-00bf-4584-a22e-f760637b52ba"). InnerVolumeSpecName "kube-api-access-pmtxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.521191 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-config-data" (OuterVolumeSpecName: "config-data") pod "e4ec4117-00bf-4584-a22e-f760637b52ba" (UID: "e4ec4117-00bf-4584-a22e-f760637b52ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.534050 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.534082 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmtxk\" (UniqueName: \"kubernetes.io/projected/e4ec4117-00bf-4584-a22e-f760637b52ba-kube-api-access-pmtxk\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.534093 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ec4117-00bf-4584-a22e-f760637b52ba-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.534102 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.567631 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4ec4117-00bf-4584-a22e-f760637b52ba" (UID: "e4ec4117-00bf-4584-a22e-f760637b52ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.588810 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e4ec4117-00bf-4584-a22e-f760637b52ba" (UID: "e4ec4117-00bf-4584-a22e-f760637b52ba"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.632059 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e4ec4117-00bf-4584-a22e-f760637b52ba" (UID: "e4ec4117-00bf-4584-a22e-f760637b52ba"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.636325 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.636346 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.636355 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ec4117-00bf-4584-a22e-f760637b52ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.760645 4735 generic.go:334] "Generic (PLEG): container finished" podID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerID="155cbd5c1a2b149ea2cfdbb545a57313f9df4e96f5035271ba7286e61de55c67" exitCode=0 Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.760718 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992","Type":"ContainerDied","Data":"155cbd5c1a2b149ea2cfdbb545a57313f9df4e96f5035271ba7286e61de55c67"} Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.762848 4735 generic.go:334] "Generic (PLEG): container finished" podID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerID="991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226" exitCode=0 Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.762909 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58678bdfb8-bgfbt" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.762924 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58678bdfb8-bgfbt" event={"ID":"e4ec4117-00bf-4584-a22e-f760637b52ba","Type":"ContainerDied","Data":"991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226"} Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.762971 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58678bdfb8-bgfbt" event={"ID":"e4ec4117-00bf-4584-a22e-f760637b52ba","Type":"ContainerDied","Data":"2b35d9cd5da1bab6bf0fe1d5421feceb001aea15433be545c81b8185a6f5a014"} Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.762994 4735 scope.go:117] "RemoveContainer" containerID="991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.765127 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.765227 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15","Type":"ContainerStarted","Data":"b0b30ff052248200e8690ad5a2aecf30d6757535d9f6c8a79edc4181fb34ee20"} Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.776794 4735 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d1cff902-6918-4f13-b1c2-cd5adad6650f" podUID="0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.798027 4735 scope.go:117] "RemoveContainer" containerID="84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.816000 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-58678bdfb8-bgfbt"] Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.835487 4735 scope.go:117] "RemoveContainer" containerID="991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226" Feb 15 20:36:12 crc kubenswrapper[4735]: E0215 20:36:12.836315 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226\": container with ID starting with 991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226 not found: ID does not exist" containerID="991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.836361 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226"} err="failed to get container status \"991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226\": rpc error: code = NotFound desc = could not find container \"991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226\": container with ID starting with 991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226 not found: ID does not exist" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.836387 4735 scope.go:117] "RemoveContainer" containerID="84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f" Feb 15 20:36:12 crc kubenswrapper[4735]: E0215 20:36:12.839591 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f\": container with ID starting with 84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f not found: ID does not exist" containerID="84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.839652 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f"} err="failed to get container status \"84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f\": rpc error: code = NotFound desc = could not find container \"84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f\": container with ID starting with 84ea545b832fa7e08db20154820991a39358df0312396f37a3bba524f11c865f not found: ID does not exist" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.879293 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-58678bdfb8-bgfbt"] Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.905145 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1cff902-6918-4f13-b1c2-cd5adad6650f" path="/var/lib/kubelet/pods/d1cff902-6918-4f13-b1c2-cd5adad6650f/volumes" Feb 15 20:36:12 crc kubenswrapper[4735]: I0215 20:36:12.906410 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4ec4117-00bf-4584-a22e-f760637b52ba" path="/var/lib/kubelet/pods/e4ec4117-00bf-4584-a22e-f760637b52ba/volumes" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.429267 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.553409 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7wl2\" (UniqueName: \"kubernetes.io/projected/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-kube-api-access-c7wl2\") pod \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.553448 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-etc-machine-id\") pod \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.553590 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data-custom\") pod \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.553675 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-scripts\") pod \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.553721 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-combined-ca-bundle\") pod \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.553744 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data\") pod \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\" (UID: \"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992\") " Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.557134 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" (UID: "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.559726 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" (UID: "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.561401 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-scripts" (OuterVolumeSpecName: "scripts") pod "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" (UID: "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.561598 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-kube-api-access-c7wl2" (OuterVolumeSpecName: "kube-api-access-c7wl2") pod "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" (UID: "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992"). InnerVolumeSpecName "kube-api-access-c7wl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.638078 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" (UID: "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.655374 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.655400 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.655408 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.655418 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7wl2\" (UniqueName: \"kubernetes.io/projected/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-kube-api-access-c7wl2\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.655431 4735 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.718213 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data" (OuterVolumeSpecName: "config-data") pod "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" (UID: "4e6f7c1d-0fd6-41f0-982e-820fdbcc6992"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.756528 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.775771 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4e6f7c1d-0fd6-41f0-982e-820fdbcc6992","Type":"ContainerDied","Data":"3e7a0514ad4979346d958611ee463ca2f852daea2529fcc93844bfc9d2cd8702"} Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.775820 4735 scope.go:117] "RemoveContainer" containerID="3125714211a6ad491c9cbe5aed193d273e3d63bd366826c59614d2d290f2ca6e" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.775905 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.823000 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.830572 4735 scope.go:117] "RemoveContainer" containerID="155cbd5c1a2b149ea2cfdbb545a57313f9df4e96f5035271ba7286e61de55c67" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.845448 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.868460 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 15 20:36:13 crc kubenswrapper[4735]: E0215 20:36:13.868809 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerName="cinder-scheduler" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.868821 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerName="cinder-scheduler" Feb 15 20:36:13 crc kubenswrapper[4735]: E0215 20:36:13.868858 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerName="placement-api" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.868864 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerName="placement-api" Feb 15 20:36:13 crc kubenswrapper[4735]: E0215 20:36:13.868878 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerName="placement-log" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.868884 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerName="placement-log" Feb 15 20:36:13 crc kubenswrapper[4735]: E0215 20:36:13.868892 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerName="probe" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.868897 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerName="probe" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.869075 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerName="cinder-scheduler" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.869089 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" containerName="probe" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.869100 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerName="placement-log" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.869110 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ec4117-00bf-4584-a22e-f760637b52ba" containerName="placement-api" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.870237 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.878571 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 15 20:36:13 crc kubenswrapper[4735]: I0215 20:36:13.887773 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.071899 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-config-data\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.071974 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7464049d-a3f2-44c0-982d-222196019aa2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.071994 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.072010 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-scripts\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.072916 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75w65\" (UniqueName: \"kubernetes.io/projected/7464049d-a3f2-44c0-982d-222196019aa2-kube-api-access-75w65\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.072993 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.174721 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75w65\" (UniqueName: \"kubernetes.io/projected/7464049d-a3f2-44c0-982d-222196019aa2-kube-api-access-75w65\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.174801 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.174980 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-config-data\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.175030 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7464049d-a3f2-44c0-982d-222196019aa2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.175093 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.175115 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-scripts\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.175109 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7464049d-a3f2-44c0-982d-222196019aa2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.181466 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.181765 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-config-data\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.188849 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-scripts\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.189283 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7464049d-a3f2-44c0-982d-222196019aa2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.202094 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75w65\" (UniqueName: \"kubernetes.io/projected/7464049d-a3f2-44c0-982d-222196019aa2-kube-api-access-75w65\") pod \"cinder-scheduler-0\" (UID: \"7464049d-a3f2-44c0-982d-222196019aa2\") " pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.219262 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.715637 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.807162 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7464049d-a3f2-44c0-982d-222196019aa2","Type":"ContainerStarted","Data":"ab9e2ef2a04cae4de152234e90d9f079094d87f7b4cd774f7ae158a5e5228226"} Feb 15 20:36:14 crc kubenswrapper[4735]: I0215 20:36:14.899374 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e6f7c1d-0fd6-41f0-982e-820fdbcc6992" path="/var/lib/kubelet/pods/4e6f7c1d-0fd6-41f0-982e-820fdbcc6992/volumes" Feb 15 20:36:15 crc kubenswrapper[4735]: W0215 20:36:15.785311 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4ec4117_00bf_4584_a22e_f760637b52ba.slice/crio-991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226.scope WatchSource:0}: Error finding container 991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226: Status 404 returned error can't find the container with id 991553270d606e1aa6f0b3f4974c753a8476d3ccedf0faf764c9b4cf7664c226 Feb 15 20:36:15 crc kubenswrapper[4735]: W0215 20:36:15.791475 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc49c8247_c8d6_4f83_9b3c_565cb6fefe51.slice/crio-f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15.scope WatchSource:0}: Error finding container f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15: Status 404 returned error can't find the container with id f712ab68b8466d9d49a2b25015c44131b2162c81c7430cd6715c73eb4659ef15 Feb 15 20:36:15 crc kubenswrapper[4735]: W0215 20:36:15.860594 4735 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e6f7c1d_0fd6_41f0_982e_820fdbcc6992.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e6f7c1d_0fd6_41f0_982e_820fdbcc6992.slice: no such file or directory Feb 15 20:36:15 crc kubenswrapper[4735]: I0215 20:36:15.887864 4735 generic.go:334] "Generic (PLEG): container finished" podID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerID="259bea590336172fa97888d69b7055d1c352e66a801b2b09b71e1977436fca5a" exitCode=137 Feb 15 20:36:15 crc kubenswrapper[4735]: I0215 20:36:15.887935 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67c7c8d94f-trlwh" event={"ID":"a20c1cd0-5f91-41de-b20d-3711518b95ef","Type":"ContainerDied","Data":"259bea590336172fa97888d69b7055d1c352e66a801b2b09b71e1977436fca5a"} Feb 15 20:36:15 crc kubenswrapper[4735]: I0215 20:36:15.892913 4735 generic.go:334] "Generic (PLEG): container finished" podID="904fa273-5436-4633-84f5-56c8205cd943" containerID="151537ee184a3d9639d7aad06d1ab88b4be09b9b675a80f059d210450a6c0d98" exitCode=137 Feb 15 20:36:15 crc kubenswrapper[4735]: I0215 20:36:15.892974 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" event={"ID":"904fa273-5436-4633-84f5-56c8205cd943","Type":"ContainerDied","Data":"151537ee184a3d9639d7aad06d1ab88b4be09b9b675a80f059d210450a6c0d98"} Feb 15 20:36:15 crc kubenswrapper[4735]: I0215 20:36:15.899611 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7464049d-a3f2-44c0-982d-222196019aa2","Type":"ContainerStarted","Data":"294d036f795c06811cd5677983296f55a8bf7557869a1e9ab7641ecb6233d613"} Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.299488 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.334652 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a20c1cd0-5f91-41de-b20d-3711518b95ef-logs\") pod \"a20c1cd0-5f91-41de-b20d-3711518b95ef\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.334755 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-combined-ca-bundle\") pod \"a20c1cd0-5f91-41de-b20d-3711518b95ef\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.334848 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data\") pod \"a20c1cd0-5f91-41de-b20d-3711518b95ef\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.334895 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data-custom\") pod \"a20c1cd0-5f91-41de-b20d-3711518b95ef\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.334974 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvhhh\" (UniqueName: \"kubernetes.io/projected/a20c1cd0-5f91-41de-b20d-3711518b95ef-kube-api-access-tvhhh\") pod \"a20c1cd0-5f91-41de-b20d-3711518b95ef\" (UID: \"a20c1cd0-5f91-41de-b20d-3711518b95ef\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.340296 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a20c1cd0-5f91-41de-b20d-3711518b95ef-logs" (OuterVolumeSpecName: "logs") pod "a20c1cd0-5f91-41de-b20d-3711518b95ef" (UID: "a20c1cd0-5f91-41de-b20d-3711518b95ef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.360111 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a20c1cd0-5f91-41de-b20d-3711518b95ef-kube-api-access-tvhhh" (OuterVolumeSpecName: "kube-api-access-tvhhh") pod "a20c1cd0-5f91-41de-b20d-3711518b95ef" (UID: "a20c1cd0-5f91-41de-b20d-3711518b95ef"). InnerVolumeSpecName "kube-api-access-tvhhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.367105 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a20c1cd0-5f91-41de-b20d-3711518b95ef" (UID: "a20c1cd0-5f91-41de-b20d-3711518b95ef"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.378312 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a20c1cd0-5f91-41de-b20d-3711518b95ef" (UID: "a20c1cd0-5f91-41de-b20d-3711518b95ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.439550 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvhhh\" (UniqueName: \"kubernetes.io/projected/a20c1cd0-5f91-41de-b20d-3711518b95ef-kube-api-access-tvhhh\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.439752 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a20c1cd0-5f91-41de-b20d-3711518b95ef-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.439836 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.439892 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.472549 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data" (OuterVolumeSpecName: "config-data") pod "a20c1cd0-5f91-41de-b20d-3711518b95ef" (UID: "a20c1cd0-5f91-41de-b20d-3711518b95ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.489693 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.541397 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-combined-ca-bundle\") pod \"904fa273-5436-4633-84f5-56c8205cd943\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.541608 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data\") pod \"904fa273-5436-4633-84f5-56c8205cd943\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.541644 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data-custom\") pod \"904fa273-5436-4633-84f5-56c8205cd943\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.541702 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/904fa273-5436-4633-84f5-56c8205cd943-logs\") pod \"904fa273-5436-4633-84f5-56c8205cd943\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.541752 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c4lg\" (UniqueName: \"kubernetes.io/projected/904fa273-5436-4633-84f5-56c8205cd943-kube-api-access-8c4lg\") pod \"904fa273-5436-4633-84f5-56c8205cd943\" (UID: \"904fa273-5436-4633-84f5-56c8205cd943\") " Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.542161 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20c1cd0-5f91-41de-b20d-3711518b95ef-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.548112 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/904fa273-5436-4633-84f5-56c8205cd943-logs" (OuterVolumeSpecName: "logs") pod "904fa273-5436-4633-84f5-56c8205cd943" (UID: "904fa273-5436-4633-84f5-56c8205cd943"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.556895 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/904fa273-5436-4633-84f5-56c8205cd943-kube-api-access-8c4lg" (OuterVolumeSpecName: "kube-api-access-8c4lg") pod "904fa273-5436-4633-84f5-56c8205cd943" (UID: "904fa273-5436-4633-84f5-56c8205cd943"). InnerVolumeSpecName "kube-api-access-8c4lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.558479 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "904fa273-5436-4633-84f5-56c8205cd943" (UID: "904fa273-5436-4633-84f5-56c8205cd943"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.606143 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data" (OuterVolumeSpecName: "config-data") pod "904fa273-5436-4633-84f5-56c8205cd943" (UID: "904fa273-5436-4633-84f5-56c8205cd943"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.611676 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "904fa273-5436-4633-84f5-56c8205cd943" (UID: "904fa273-5436-4633-84f5-56c8205cd943"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.643958 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.643989 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.644022 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/904fa273-5436-4633-84f5-56c8205cd943-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.644033 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c4lg\" (UniqueName: \"kubernetes.io/projected/904fa273-5436-4633-84f5-56c8205cd943-kube-api-access-8c4lg\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.644043 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/904fa273-5436-4633-84f5-56c8205cd943-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.917634 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67c7c8d94f-trlwh" event={"ID":"a20c1cd0-5f91-41de-b20d-3711518b95ef","Type":"ContainerDied","Data":"c340ab8e64181fba0ff6704d6f0e337cde1acf656755cf8cb751b453aabbe204"} Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.917684 4735 scope.go:117] "RemoveContainer" containerID="259bea590336172fa97888d69b7055d1c352e66a801b2b09b71e1977436fca5a" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.918247 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-67c7c8d94f-trlwh" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.924790 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" event={"ID":"904fa273-5436-4633-84f5-56c8205cd943","Type":"ContainerDied","Data":"fd36fac619467d9278527ca97669feaa20216178123d0981b2c4364d32cf44fc"} Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.924872 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-684fc99c4-wchtt" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.937153 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7464049d-a3f2-44c0-982d-222196019aa2","Type":"ContainerStarted","Data":"821c7da26165db602b333873f4fb4410279ea9f0b3bcebd5f15e3d397ba903d7"} Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.944346 4735 scope.go:117] "RemoveContainer" containerID="eefa0ed78309417ca1db111a3d8580f1c3369b7be5c327884253132a62d51333" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.951330 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-67c7c8d94f-trlwh"] Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.965534 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-67c7c8d94f-trlwh"] Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.966310 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.966299503 podStartE2EDuration="3.966299503s" podCreationTimestamp="2026-02-15 20:36:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:16.960564928 +0000 UTC m=+1184.826580551" watchObservedRunningTime="2026-02-15 20:36:16.966299503 +0000 UTC m=+1184.832315116" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.971702 4735 scope.go:117] "RemoveContainer" containerID="151537ee184a3d9639d7aad06d1ab88b4be09b9b675a80f059d210450a6c0d98" Feb 15 20:36:16 crc kubenswrapper[4735]: I0215 20:36:16.993054 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-684fc99c4-wchtt"] Feb 15 20:36:17 crc kubenswrapper[4735]: I0215 20:36:17.000107 4735 scope.go:117] "RemoveContainer" containerID="4a8cd143180cf340ef0b1a459c8443020429f5971a4b804256913003887dc484" Feb 15 20:36:17 crc kubenswrapper[4735]: I0215 20:36:17.000275 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-684fc99c4-wchtt"] Feb 15 20:36:18 crc kubenswrapper[4735]: I0215 20:36:18.724238 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-fdf549c7-xpft7" Feb 15 20:36:18 crc kubenswrapper[4735]: I0215 20:36:18.802880 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-59f96fd876-q78bj"] Feb 15 20:36:18 crc kubenswrapper[4735]: I0215 20:36:18.803476 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-59f96fd876-q78bj" podUID="4e551f68-1918-4b4a-99de-c7e05c507187" containerName="neutron-httpd" containerID="cri-o://51c478a3412d9dd59ac3624eb9e78d9ded5143fcdf90a0f76ed2fc9271a76bf1" gracePeriod=30 Feb 15 20:36:18 crc kubenswrapper[4735]: I0215 20:36:18.809225 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-59f96fd876-q78bj" podUID="4e551f68-1918-4b4a-99de-c7e05c507187" containerName="neutron-api" containerID="cri-o://7c06fc1659d45c78cd1d29f4b44200f8c4675dc45011376308cfbc0b5abfbba7" gracePeriod=30 Feb 15 20:36:18 crc kubenswrapper[4735]: I0215 20:36:18.908032 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="904fa273-5436-4633-84f5-56c8205cd943" path="/var/lib/kubelet/pods/904fa273-5436-4633-84f5-56c8205cd943/volumes" Feb 15 20:36:18 crc kubenswrapper[4735]: I0215 20:36:18.908703 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a20c1cd0-5f91-41de-b20d-3711518b95ef" path="/var/lib/kubelet/pods/a20c1cd0-5f91-41de-b20d-3711518b95ef/volumes" Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.222054 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.679780 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.680096 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.680138 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.680770 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"94228183c3d1abcbd5b788cf2f92ec757e11f8d18a65dfdd44912dcfc7991e88"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.680825 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://94228183c3d1abcbd5b788cf2f92ec757e11f8d18a65dfdd44912dcfc7991e88" gracePeriod=600 Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.985641 4735 generic.go:334] "Generic (PLEG): container finished" podID="4e551f68-1918-4b4a-99de-c7e05c507187" containerID="51c478a3412d9dd59ac3624eb9e78d9ded5143fcdf90a0f76ed2fc9271a76bf1" exitCode=0 Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.985698 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59f96fd876-q78bj" event={"ID":"4e551f68-1918-4b4a-99de-c7e05c507187","Type":"ContainerDied","Data":"51c478a3412d9dd59ac3624eb9e78d9ded5143fcdf90a0f76ed2fc9271a76bf1"} Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.989795 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="94228183c3d1abcbd5b788cf2f92ec757e11f8d18a65dfdd44912dcfc7991e88" exitCode=0 Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.989824 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"94228183c3d1abcbd5b788cf2f92ec757e11f8d18a65dfdd44912dcfc7991e88"} Feb 15 20:36:19 crc kubenswrapper[4735]: I0215 20:36:19.989849 4735 scope.go:117] "RemoveContainer" containerID="776a88a7dd582588d5537251275510c6e25c50e714d7281e0f68c913b0f4fe42" Feb 15 20:36:20 crc kubenswrapper[4735]: I0215 20:36:20.389964 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:36:20 crc kubenswrapper[4735]: I0215 20:36:20.390202 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerName="glance-log" containerID="cri-o://0eceef46b722125a7a67a2e0b988daa2f542ecc7fd172b767ea9348df61f5f01" gracePeriod=30 Feb 15 20:36:20 crc kubenswrapper[4735]: I0215 20:36:20.390289 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerName="glance-httpd" containerID="cri-o://2362903ea75605e58912d9ab9e9910e689415b6e37485ede12f5837307f8d68a" gracePeriod=30 Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.047007 4735 generic.go:334] "Generic (PLEG): container finished" podID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerID="0eceef46b722125a7a67a2e0b988daa2f542ecc7fd172b767ea9348df61f5f01" exitCode=143 Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.047413 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bdadc81-14d1-469b-82c1-68f79cfa71a9","Type":"ContainerDied","Data":"0eceef46b722125a7a67a2e0b988daa2f542ecc7fd172b767ea9348df61f5f01"} Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.057094 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-b7b86987-zpgc8"] Feb 15 20:36:21 crc kubenswrapper[4735]: E0215 20:36:21.057812 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerName="barbican-worker" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.057828 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerName="barbican-worker" Feb 15 20:36:21 crc kubenswrapper[4735]: E0215 20:36:21.057845 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="904fa273-5436-4633-84f5-56c8205cd943" containerName="barbican-keystone-listener-log" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.057853 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="904fa273-5436-4633-84f5-56c8205cd943" containerName="barbican-keystone-listener-log" Feb 15 20:36:21 crc kubenswrapper[4735]: E0215 20:36:21.057883 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="904fa273-5436-4633-84f5-56c8205cd943" containerName="barbican-keystone-listener" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.057893 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="904fa273-5436-4633-84f5-56c8205cd943" containerName="barbican-keystone-listener" Feb 15 20:36:21 crc kubenswrapper[4735]: E0215 20:36:21.057906 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerName="barbican-worker-log" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.057912 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerName="barbican-worker-log" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.058255 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="904fa273-5436-4633-84f5-56c8205cd943" containerName="barbican-keystone-listener" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.058283 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="904fa273-5436-4633-84f5-56c8205cd943" containerName="barbican-keystone-listener-log" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.058305 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerName="barbican-worker-log" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.058318 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a20c1cd0-5f91-41de-b20d-3711518b95ef" containerName="barbican-worker" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.059722 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.063295 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"97239e5cb100e71e3f44686a5d1c9c6d1a3b14c88d00878cdd46ec7dbf229a42"} Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.071007 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.071979 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.072309 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.088349 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-b7b86987-zpgc8"] Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.241104 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/843b8077-b23a-42f9-bf70-79767ce35b4f-etc-swift\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.241434 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-public-tls-certs\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.241486 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/843b8077-b23a-42f9-bf70-79767ce35b4f-log-httpd\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.241501 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/843b8077-b23a-42f9-bf70-79767ce35b4f-run-httpd\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.241520 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-config-data\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.241535 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-internal-tls-certs\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.241563 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-combined-ca-bundle\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.241633 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg9pz\" (UniqueName: \"kubernetes.io/projected/843b8077-b23a-42f9-bf70-79767ce35b4f-kube-api-access-wg9pz\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.343208 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/843b8077-b23a-42f9-bf70-79767ce35b4f-etc-swift\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.343253 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-public-tls-certs\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.343291 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/843b8077-b23a-42f9-bf70-79767ce35b4f-log-httpd\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.343328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/843b8077-b23a-42f9-bf70-79767ce35b4f-run-httpd\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.343354 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-config-data\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.343383 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-internal-tls-certs\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.343413 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-combined-ca-bundle\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.343483 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg9pz\" (UniqueName: \"kubernetes.io/projected/843b8077-b23a-42f9-bf70-79767ce35b4f-kube-api-access-wg9pz\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.345869 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/843b8077-b23a-42f9-bf70-79767ce35b4f-log-httpd\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.352182 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/843b8077-b23a-42f9-bf70-79767ce35b4f-run-httpd\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.354404 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-internal-tls-certs\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.356769 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/843b8077-b23a-42f9-bf70-79767ce35b4f-etc-swift\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.363975 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-config-data\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.364718 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-public-tls-certs\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.371205 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843b8077-b23a-42f9-bf70-79767ce35b4f-combined-ca-bundle\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.372081 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg9pz\" (UniqueName: \"kubernetes.io/projected/843b8077-b23a-42f9-bf70-79767ce35b4f-kube-api-access-wg9pz\") pod \"swift-proxy-b7b86987-zpgc8\" (UID: \"843b8077-b23a-42f9-bf70-79767ce35b4f\") " pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:21 crc kubenswrapper[4735]: I0215 20:36:21.402180 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:22 crc kubenswrapper[4735]: E0215 20:36:22.151430 4735 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/be56e9ef7a92008b2dda3654304d634d31b75f85e566594c6edd7ff951be88e1/diff" to get inode usage: stat /var/lib/containers/storage/overlay/be56e9ef7a92008b2dda3654304d634d31b75f85e566594c6edd7ff951be88e1/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_5d7a00e9-2271-49b1-a74f-51f5f668995a/ceilometer-notification-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_5d7a00e9-2271-49b1-a74f-51f5f668995a/ceilometer-notification-agent/0.log: no such file or directory Feb 15 20:36:22 crc kubenswrapper[4735]: I0215 20:36:22.863362 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:22 crc kubenswrapper[4735]: I0215 20:36:22.865775 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="ceilometer-central-agent" containerID="cri-o://8d114b6173c49bd9cb2de87ba812abf6cde86c1ef7d83bcbe0aaff2a24425584" gracePeriod=30 Feb 15 20:36:22 crc kubenswrapper[4735]: I0215 20:36:22.866582 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="proxy-httpd" containerID="cri-o://73e5563131beea3e291af24d55b00d086bcd2089890020e6c943154f24febfce" gracePeriod=30 Feb 15 20:36:22 crc kubenswrapper[4735]: I0215 20:36:22.866639 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="sg-core" containerID="cri-o://6c226ff20ed0038cef94ee50f1214e3a0e63d9739012d1195ec9cb4b02b520c1" gracePeriod=30 Feb 15 20:36:22 crc kubenswrapper[4735]: I0215 20:36:22.866670 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="ceilometer-notification-agent" containerID="cri-o://563cd8683e20b062dc2adde56b23ee47afd98fa9e99b53d61e87b491b24d6835" gracePeriod=30 Feb 15 20:36:22 crc kubenswrapper[4735]: I0215 20:36:22.884131 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.172:3000/\": EOF" Feb 15 20:36:23 crc kubenswrapper[4735]: I0215 20:36:23.083915 4735 generic.go:334] "Generic (PLEG): container finished" podID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerID="6c226ff20ed0038cef94ee50f1214e3a0e63d9739012d1195ec9cb4b02b520c1" exitCode=2 Feb 15 20:36:23 crc kubenswrapper[4735]: I0215 20:36:23.083967 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerDied","Data":"6c226ff20ed0038cef94ee50f1214e3a0e63d9739012d1195ec9cb4b02b520c1"} Feb 15 20:36:24 crc kubenswrapper[4735]: I0215 20:36:24.096817 4735 generic.go:334] "Generic (PLEG): container finished" podID="4e551f68-1918-4b4a-99de-c7e05c507187" containerID="7c06fc1659d45c78cd1d29f4b44200f8c4675dc45011376308cfbc0b5abfbba7" exitCode=0 Feb 15 20:36:24 crc kubenswrapper[4735]: I0215 20:36:24.096903 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59f96fd876-q78bj" event={"ID":"4e551f68-1918-4b4a-99de-c7e05c507187","Type":"ContainerDied","Data":"7c06fc1659d45c78cd1d29f4b44200f8c4675dc45011376308cfbc0b5abfbba7"} Feb 15 20:36:24 crc kubenswrapper[4735]: I0215 20:36:24.104271 4735 generic.go:334] "Generic (PLEG): container finished" podID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerID="73e5563131beea3e291af24d55b00d086bcd2089890020e6c943154f24febfce" exitCode=0 Feb 15 20:36:24 crc kubenswrapper[4735]: I0215 20:36:24.104304 4735 generic.go:334] "Generic (PLEG): container finished" podID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerID="8d114b6173c49bd9cb2de87ba812abf6cde86c1ef7d83bcbe0aaff2a24425584" exitCode=0 Feb 15 20:36:24 crc kubenswrapper[4735]: I0215 20:36:24.104358 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerDied","Data":"73e5563131beea3e291af24d55b00d086bcd2089890020e6c943154f24febfce"} Feb 15 20:36:24 crc kubenswrapper[4735]: I0215 20:36:24.104384 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerDied","Data":"8d114b6173c49bd9cb2de87ba812abf6cde86c1ef7d83bcbe0aaff2a24425584"} Feb 15 20:36:24 crc kubenswrapper[4735]: I0215 20:36:24.107469 4735 generic.go:334] "Generic (PLEG): container finished" podID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerID="2362903ea75605e58912d9ab9e9910e689415b6e37485ede12f5837307f8d68a" exitCode=0 Feb 15 20:36:24 crc kubenswrapper[4735]: I0215 20:36:24.107513 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bdadc81-14d1-469b-82c1-68f79cfa71a9","Type":"ContainerDied","Data":"2362903ea75605e58912d9ab9e9910e689415b6e37485ede12f5837307f8d68a"} Feb 15 20:36:24 crc kubenswrapper[4735]: I0215 20:36:24.540261 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 15 20:36:25 crc kubenswrapper[4735]: I0215 20:36:25.137077 4735 generic.go:334] "Generic (PLEG): container finished" podID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerID="509c549dfe02bd37ca623752f3760afb9e5b4c41da36b238f99fb78d66aac256" exitCode=137 Feb 15 20:36:25 crc kubenswrapper[4735]: I0215 20:36:25.137165 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerDied","Data":"509c549dfe02bd37ca623752f3760afb9e5b4c41da36b238f99fb78d66aac256"} Feb 15 20:36:25 crc kubenswrapper[4735]: I0215 20:36:25.141251 4735 generic.go:334] "Generic (PLEG): container finished" podID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerID="063addf10589b9afd3c588e4171de6d504c19a3cbf44f266f86a25dfe5ca99e3" exitCode=137 Feb 15 20:36:25 crc kubenswrapper[4735]: I0215 20:36:25.141285 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c647fbf5b-7zs9r" event={"ID":"970d5293-b621-45e1-a1e4-8cc176c9a148","Type":"ContainerDied","Data":"063addf10589b9afd3c588e4171de6d504c19a3cbf44f266f86a25dfe5ca99e3"} Feb 15 20:36:26 crc kubenswrapper[4735]: I0215 20:36:26.115445 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:36:26 crc kubenswrapper[4735]: I0215 20:36:26.115879 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerName="glance-log" containerID="cri-o://7f220f4630db2ac2af4c43eee7e69a114fc9d4b048e5ff25b90aaa27e470205d" gracePeriod=30 Feb 15 20:36:26 crc kubenswrapper[4735]: I0215 20:36:26.115981 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerName="glance-httpd" containerID="cri-o://fce032895e4ab9a42d9c005265ba71abf97dc38fd3a315afabcc1cd553214a3e" gracePeriod=30 Feb 15 20:36:26 crc kubenswrapper[4735]: I0215 20:36:26.150726 4735 generic.go:334] "Generic (PLEG): container finished" podID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerID="563cd8683e20b062dc2adde56b23ee47afd98fa9e99b53d61e87b491b24d6835" exitCode=0 Feb 15 20:36:26 crc kubenswrapper[4735]: I0215 20:36:26.150766 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerDied","Data":"563cd8683e20b062dc2adde56b23ee47afd98fa9e99b53d61e87b491b24d6835"} Feb 15 20:36:27 crc kubenswrapper[4735]: I0215 20:36:27.160905 4735 generic.go:334] "Generic (PLEG): container finished" podID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerID="7f220f4630db2ac2af4c43eee7e69a114fc9d4b048e5ff25b90aaa27e470205d" exitCode=143 Feb 15 20:36:27 crc kubenswrapper[4735]: I0215 20:36:27.160967 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e48a6f8b-a91e-438b-832b-b9ba15c87e15","Type":"ContainerDied","Data":"7f220f4630db2ac2af4c43eee7e69a114fc9d4b048e5ff25b90aaa27e470205d"} Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.040518 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-svbdx"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.041774 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.090267 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-svbdx"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.096697 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvz4w\" (UniqueName: \"kubernetes.io/projected/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-kube-api-access-mvz4w\") pod \"nova-api-db-create-svbdx\" (UID: \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\") " pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.096768 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-operator-scripts\") pod \"nova-api-db-create-svbdx\" (UID: \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\") " pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.198012 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvz4w\" (UniqueName: \"kubernetes.io/projected/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-kube-api-access-mvz4w\") pod \"nova-api-db-create-svbdx\" (UID: \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\") " pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.198085 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-operator-scripts\") pod \"nova-api-db-create-svbdx\" (UID: \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\") " pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.198885 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-operator-scripts\") pod \"nova-api-db-create-svbdx\" (UID: \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\") " pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.228644 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvz4w\" (UniqueName: \"kubernetes.io/projected/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-kube-api-access-mvz4w\") pod \"nova-api-db-create-svbdx\" (UID: \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\") " pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.241012 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-f66hh"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.242526 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.245914 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-f66hh"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.305630 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8kzf\" (UniqueName: \"kubernetes.io/projected/85484764-8eab-4bbd-8d2f-1726f287ada3-kube-api-access-d8kzf\") pod \"nova-cell0-db-create-f66hh\" (UID: \"85484764-8eab-4bbd-8d2f-1726f287ada3\") " pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.306228 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85484764-8eab-4bbd-8d2f-1726f287ada3-operator-scripts\") pod \"nova-cell0-db-create-f66hh\" (UID: \"85484764-8eab-4bbd-8d2f-1726f287ada3\") " pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.358369 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.404145 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-af86-account-create-update-5mp2z"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.407879 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8kzf\" (UniqueName: \"kubernetes.io/projected/85484764-8eab-4bbd-8d2f-1726f287ada3-kube-api-access-d8kzf\") pod \"nova-cell0-db-create-f66hh\" (UID: \"85484764-8eab-4bbd-8d2f-1726f287ada3\") " pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.407957 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85484764-8eab-4bbd-8d2f-1726f287ada3-operator-scripts\") pod \"nova-cell0-db-create-f66hh\" (UID: \"85484764-8eab-4bbd-8d2f-1726f287ada3\") " pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.408685 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85484764-8eab-4bbd-8d2f-1726f287ada3-operator-scripts\") pod \"nova-cell0-db-create-f66hh\" (UID: \"85484764-8eab-4bbd-8d2f-1726f287ada3\") " pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.413353 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.423716 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.439301 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-af86-account-create-update-5mp2z"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.467625 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8kzf\" (UniqueName: \"kubernetes.io/projected/85484764-8eab-4bbd-8d2f-1726f287ada3-kube-api-access-d8kzf\") pod \"nova-cell0-db-create-f66hh\" (UID: \"85484764-8eab-4bbd-8d2f-1726f287ada3\") " pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.503022 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-m6ztx"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.504216 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.511104 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27472c1-f67e-49e6-bd56-a36101f69c64-operator-scripts\") pod \"nova-api-af86-account-create-update-5mp2z\" (UID: \"f27472c1-f67e-49e6-bd56-a36101f69c64\") " pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.511167 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48189563-f3c7-49a0-b01e-86b25b077fd5-operator-scripts\") pod \"nova-cell1-db-create-m6ztx\" (UID: \"48189563-f3c7-49a0-b01e-86b25b077fd5\") " pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.511197 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29x5h\" (UniqueName: \"kubernetes.io/projected/f27472c1-f67e-49e6-bd56-a36101f69c64-kube-api-access-29x5h\") pod \"nova-api-af86-account-create-update-5mp2z\" (UID: \"f27472c1-f67e-49e6-bd56-a36101f69c64\") " pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.511339 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v99f\" (UniqueName: \"kubernetes.io/projected/48189563-f3c7-49a0-b01e-86b25b077fd5-kube-api-access-8v99f\") pod \"nova-cell1-db-create-m6ztx\" (UID: \"48189563-f3c7-49a0-b01e-86b25b077fd5\") " pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.541026 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-m6ztx"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.625441 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v99f\" (UniqueName: \"kubernetes.io/projected/48189563-f3c7-49a0-b01e-86b25b077fd5-kube-api-access-8v99f\") pod \"nova-cell1-db-create-m6ztx\" (UID: \"48189563-f3c7-49a0-b01e-86b25b077fd5\") " pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.625508 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27472c1-f67e-49e6-bd56-a36101f69c64-operator-scripts\") pod \"nova-api-af86-account-create-update-5mp2z\" (UID: \"f27472c1-f67e-49e6-bd56-a36101f69c64\") " pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.625542 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48189563-f3c7-49a0-b01e-86b25b077fd5-operator-scripts\") pod \"nova-cell1-db-create-m6ztx\" (UID: \"48189563-f3c7-49a0-b01e-86b25b077fd5\") " pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.625563 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29x5h\" (UniqueName: \"kubernetes.io/projected/f27472c1-f67e-49e6-bd56-a36101f69c64-kube-api-access-29x5h\") pod \"nova-api-af86-account-create-update-5mp2z\" (UID: \"f27472c1-f67e-49e6-bd56-a36101f69c64\") " pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.626495 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27472c1-f67e-49e6-bd56-a36101f69c64-operator-scripts\") pod \"nova-api-af86-account-create-update-5mp2z\" (UID: \"f27472c1-f67e-49e6-bd56-a36101f69c64\") " pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.626534 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48189563-f3c7-49a0-b01e-86b25b077fd5-operator-scripts\") pod \"nova-cell1-db-create-m6ztx\" (UID: \"48189563-f3c7-49a0-b01e-86b25b077fd5\") " pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.632207 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.650507 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v99f\" (UniqueName: \"kubernetes.io/projected/48189563-f3c7-49a0-b01e-86b25b077fd5-kube-api-access-8v99f\") pod \"nova-cell1-db-create-m6ztx\" (UID: \"48189563-f3c7-49a0-b01e-86b25b077fd5\") " pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.657461 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29x5h\" (UniqueName: \"kubernetes.io/projected/f27472c1-f67e-49e6-bd56-a36101f69c64-kube-api-access-29x5h\") pod \"nova-api-af86-account-create-update-5mp2z\" (UID: \"f27472c1-f67e-49e6-bd56-a36101f69c64\") " pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.745099 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-33f9-account-create-update-t72d7"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.748157 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.751990 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.768297 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-33f9-account-create-update-t72d7"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.793534 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.834615 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-d824-account-create-update-gmj74"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.836597 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.842177 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.842742 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/493558c5-9375-43a2-92e1-9f20470fc681-operator-scripts\") pod \"nova-cell0-33f9-account-create-update-t72d7\" (UID: \"493558c5-9375-43a2-92e1-9f20470fc681\") " pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.851293 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2h9v\" (UniqueName: \"kubernetes.io/projected/493558c5-9375-43a2-92e1-9f20470fc681-kube-api-access-z2h9v\") pod \"nova-cell0-33f9-account-create-update-t72d7\" (UID: \"493558c5-9375-43a2-92e1-9f20470fc681\") " pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.855254 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.879693 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d824-account-create-update-gmj74"] Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.960994 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2h9v\" (UniqueName: \"kubernetes.io/projected/493558c5-9375-43a2-92e1-9f20470fc681-kube-api-access-z2h9v\") pod \"nova-cell0-33f9-account-create-update-t72d7\" (UID: \"493558c5-9375-43a2-92e1-9f20470fc681\") " pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.961214 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2d80b54-e0f6-415a-9bab-1cff8e208efa-operator-scripts\") pod \"nova-cell1-d824-account-create-update-gmj74\" (UID: \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\") " pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.961238 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dsk9\" (UniqueName: \"kubernetes.io/projected/f2d80b54-e0f6-415a-9bab-1cff8e208efa-kube-api-access-4dsk9\") pod \"nova-cell1-d824-account-create-update-gmj74\" (UID: \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\") " pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.961275 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/493558c5-9375-43a2-92e1-9f20470fc681-operator-scripts\") pod \"nova-cell0-33f9-account-create-update-t72d7\" (UID: \"493558c5-9375-43a2-92e1-9f20470fc681\") " pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.962003 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/493558c5-9375-43a2-92e1-9f20470fc681-operator-scripts\") pod \"nova-cell0-33f9-account-create-update-t72d7\" (UID: \"493558c5-9375-43a2-92e1-9f20470fc681\") " pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:29 crc kubenswrapper[4735]: I0215 20:36:29.985007 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2h9v\" (UniqueName: \"kubernetes.io/projected/493558c5-9375-43a2-92e1-9f20470fc681-kube-api-access-z2h9v\") pod \"nova-cell0-33f9-account-create-update-t72d7\" (UID: \"493558c5-9375-43a2-92e1-9f20470fc681\") " pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.066438 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2d80b54-e0f6-415a-9bab-1cff8e208efa-operator-scripts\") pod \"nova-cell1-d824-account-create-update-gmj74\" (UID: \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\") " pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.066488 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dsk9\" (UniqueName: \"kubernetes.io/projected/f2d80b54-e0f6-415a-9bab-1cff8e208efa-kube-api-access-4dsk9\") pod \"nova-cell1-d824-account-create-update-gmj74\" (UID: \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\") " pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.067807 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2d80b54-e0f6-415a-9bab-1cff8e208efa-operator-scripts\") pod \"nova-cell1-d824-account-create-update-gmj74\" (UID: \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\") " pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.094525 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dsk9\" (UniqueName: \"kubernetes.io/projected/f2d80b54-e0f6-415a-9bab-1cff8e208efa-kube-api-access-4dsk9\") pod \"nova-cell1-d824-account-create-update-gmj74\" (UID: \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\") " pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.152050 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.229915 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.318854 4735 generic.go:334] "Generic (PLEG): container finished" podID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerID="fce032895e4ab9a42d9c005265ba71abf97dc38fd3a315afabcc1cd553214a3e" exitCode=0 Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.319166 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e48a6f8b-a91e-438b-832b-b9ba15c87e15","Type":"ContainerDied","Data":"fce032895e4ab9a42d9c005265ba71abf97dc38fd3a315afabcc1cd553214a3e"} Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.364210 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerStarted","Data":"86ffd054acc51bbf8b9d87f36ab9440eeb9c687baf017ae70a357fa95727f187"} Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.403628 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c647fbf5b-7zs9r" event={"ID":"970d5293-b621-45e1-a1e4-8cc176c9a148","Type":"ContainerStarted","Data":"dfabab6e6193ecc85b9a47a5ec99565c4c3f90470cab4a3b43a98447f3711f35"} Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.494181 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.540923 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703309 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-config\") pod \"4e551f68-1918-4b4a-99de-c7e05c507187\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703656 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-run-httpd\") pod \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703698 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-scripts\") pod \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703755 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xml2t\" (UniqueName: \"kubernetes.io/projected/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-kube-api-access-xml2t\") pod \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703799 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-combined-ca-bundle\") pod \"4e551f68-1918-4b4a-99de-c7e05c507187\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703825 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-log-httpd\") pod \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703851 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-httpd-config\") pod \"4e551f68-1918-4b4a-99de-c7e05c507187\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703873 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-config-data\") pod \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703925 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-combined-ca-bundle\") pod \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.703990 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-sg-core-conf-yaml\") pod \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\" (UID: \"49a00bb1-fbaf-45a7-882b-6a2a24bd100d\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.704021 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-ovndb-tls-certs\") pod \"4e551f68-1918-4b4a-99de-c7e05c507187\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.704088 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl48k\" (UniqueName: \"kubernetes.io/projected/4e551f68-1918-4b4a-99de-c7e05c507187-kube-api-access-vl48k\") pod \"4e551f68-1918-4b4a-99de-c7e05c507187\" (UID: \"4e551f68-1918-4b4a-99de-c7e05c507187\") " Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.710934 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "49a00bb1-fbaf-45a7-882b-6a2a24bd100d" (UID: "49a00bb1-fbaf-45a7-882b-6a2a24bd100d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.715726 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "49a00bb1-fbaf-45a7-882b-6a2a24bd100d" (UID: "49a00bb1-fbaf-45a7-882b-6a2a24bd100d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:30 crc kubenswrapper[4735]: W0215 20:36:30.737157 4735 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1cff902_6918_4f13_b1c2_cd5adad6650f.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1cff902_6918_4f13_b1c2_cd5adad6650f.slice: no such file or directory Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.750156 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4e551f68-1918-4b4a-99de-c7e05c507187" (UID: "4e551f68-1918-4b4a-99de-c7e05c507187"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.771148 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-kube-api-access-xml2t" (OuterVolumeSpecName: "kube-api-access-xml2t") pod "49a00bb1-fbaf-45a7-882b-6a2a24bd100d" (UID: "49a00bb1-fbaf-45a7-882b-6a2a24bd100d"). InnerVolumeSpecName "kube-api-access-xml2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.787719 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e551f68-1918-4b4a-99de-c7e05c507187-kube-api-access-vl48k" (OuterVolumeSpecName: "kube-api-access-vl48k") pod "4e551f68-1918-4b4a-99de-c7e05c507187" (UID: "4e551f68-1918-4b4a-99de-c7e05c507187"). InnerVolumeSpecName "kube-api-access-vl48k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.806477 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-scripts" (OuterVolumeSpecName: "scripts") pod "49a00bb1-fbaf-45a7-882b-6a2a24bd100d" (UID: "49a00bb1-fbaf-45a7-882b-6a2a24bd100d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.811845 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.811874 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.811883 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xml2t\" (UniqueName: \"kubernetes.io/projected/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-kube-api-access-xml2t\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.811892 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.811902 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.811910 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl48k\" (UniqueName: \"kubernetes.io/projected/4e551f68-1918-4b4a-99de-c7e05c507187-kube-api-access-vl48k\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.881899 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:30 crc kubenswrapper[4735]: I0215 20:36:30.996311 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-config" (OuterVolumeSpecName: "config") pod "4e551f68-1918-4b4a-99de-c7e05c507187" (UID: "4e551f68-1918-4b4a-99de-c7e05c507187"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.010880 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpj27\" (UniqueName: \"kubernetes.io/projected/1bdadc81-14d1-469b-82c1-68f79cfa71a9-kube-api-access-qpj27\") pod \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.010930 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-config-data\") pod \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.011105 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-httpd-run\") pod \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.011158 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.011228 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-combined-ca-bundle\") pod \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.011308 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-scripts\") pod \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.011330 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-internal-tls-certs\") pod \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.011354 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-logs\") pod \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\" (UID: \"1bdadc81-14d1-469b-82c1-68f79cfa71a9\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.015668 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.021074 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "49a00bb1-fbaf-45a7-882b-6a2a24bd100d" (UID: "49a00bb1-fbaf-45a7-882b-6a2a24bd100d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.024098 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bdadc81-14d1-469b-82c1-68f79cfa71a9-kube-api-access-qpj27" (OuterVolumeSpecName: "kube-api-access-qpj27") pod "1bdadc81-14d1-469b-82c1-68f79cfa71a9" (UID: "1bdadc81-14d1-469b-82c1-68f79cfa71a9"). InnerVolumeSpecName "kube-api-access-qpj27". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.028042 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1bdadc81-14d1-469b-82c1-68f79cfa71a9" (UID: "1bdadc81-14d1-469b-82c1-68f79cfa71a9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.030365 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-logs" (OuterVolumeSpecName: "logs") pod "1bdadc81-14d1-469b-82c1-68f79cfa71a9" (UID: "1bdadc81-14d1-469b-82c1-68f79cfa71a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.039920 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e551f68-1918-4b4a-99de-c7e05c507187" (UID: "4e551f68-1918-4b4a-99de-c7e05c507187"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.041059 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-scripts" (OuterVolumeSpecName: "scripts") pod "1bdadc81-14d1-469b-82c1-68f79cfa71a9" (UID: "1bdadc81-14d1-469b-82c1-68f79cfa71a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.041516 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "1bdadc81-14d1-469b-82c1-68f79cfa71a9" (UID: "1bdadc81-14d1-469b-82c1-68f79cfa71a9"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.134168 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpj27\" (UniqueName: \"kubernetes.io/projected/1bdadc81-14d1-469b-82c1-68f79cfa71a9-kube-api-access-qpj27\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.134199 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.134214 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.134237 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.134250 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.134259 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.134268 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bdadc81-14d1-469b-82c1-68f79cfa71a9-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.170478 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-svbdx"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.233981 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4e551f68-1918-4b4a-99de-c7e05c507187" (UID: "4e551f68-1918-4b4a-99de-c7e05c507187"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.240366 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bdadc81-14d1-469b-82c1-68f79cfa71a9" (UID: "1bdadc81-14d1-469b-82c1-68f79cfa71a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.264736 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.272508 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-b7b86987-zpgc8"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.275357 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.275382 4735 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e551f68-1918-4b4a-99de-c7e05c507187-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.275393 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.323515 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1bdadc81-14d1-469b-82c1-68f79cfa71a9" (UID: "1bdadc81-14d1-469b-82c1-68f79cfa71a9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.359532 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-config-data" (OuterVolumeSpecName: "config-data") pod "49a00bb1-fbaf-45a7-882b-6a2a24bd100d" (UID: "49a00bb1-fbaf-45a7-882b-6a2a24bd100d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.372222 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-config-data" (OuterVolumeSpecName: "config-data") pod "1bdadc81-14d1-469b-82c1-68f79cfa71a9" (UID: "1bdadc81-14d1-469b-82c1-68f79cfa71a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.377231 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.377266 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdadc81-14d1-469b-82c1-68f79cfa71a9-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.377278 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.389447 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49a00bb1-fbaf-45a7-882b-6a2a24bd100d" (UID: "49a00bb1-fbaf-45a7-882b-6a2a24bd100d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.440653 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bdadc81-14d1-469b-82c1-68f79cfa71a9","Type":"ContainerDied","Data":"55361a0c531f4e64c8e96845063e9a885a0d31dc1a08b49bbe8b89f8c874ea89"} Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.440701 4735 scope.go:117] "RemoveContainer" containerID="2362903ea75605e58912d9ab9e9910e689415b6e37485ede12f5837307f8d68a" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.440837 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.447357 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.453354 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15","Type":"ContainerStarted","Data":"52cd42ac9aeb2877191b97fc6e4024572c3f33f318717f1d60b9c9da4a9ffef4"} Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.463126 4735 generic.go:334] "Generic (PLEG): container finished" podID="107aa5db-f0d0-4817-8d2d-12da89492473" containerID="be1dc831e7060f2c2a416d0785c4bfb76e59d77ebfaab5dd8be8cff4c319dd71" exitCode=137 Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.463402 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"107aa5db-f0d0-4817-8d2d-12da89492473","Type":"ContainerDied","Data":"be1dc831e7060f2c2a416d0785c4bfb76e59d77ebfaab5dd8be8cff4c319dd71"} Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.478769 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a00bb1-fbaf-45a7-882b-6a2a24bd100d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.508316 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-svbdx" event={"ID":"c8507ec7-d881-4a4a-82dc-3af56d4d53aa","Type":"ContainerStarted","Data":"47839f57575740c82944bf7680f988009debb423027efe1dfb3d6c662e104530"} Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.528312 4735 scope.go:117] "RemoveContainer" containerID="0eceef46b722125a7a67a2e0b988daa2f542ecc7fd172b767ea9348df61f5f01" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.545135 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59f96fd876-q78bj" event={"ID":"4e551f68-1918-4b4a-99de-c7e05c507187","Type":"ContainerDied","Data":"ce1d49f8915c985e900f99b17ba2c54639fa8577c4f44b083ef75d73f58294d4"} Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.545350 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59f96fd876-q78bj" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.557001 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.584751 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a00bb1-fbaf-45a7-882b-6a2a24bd100d","Type":"ContainerDied","Data":"d81b7261a3462028a97a77e2a7b77d8f13094f5d866f42e0cf538bd7412e1013"} Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.584910 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.585236 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-combined-ca-bundle\") pod \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.585501 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-public-tls-certs\") pod \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.585525 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk4zs\" (UniqueName: \"kubernetes.io/projected/e48a6f8b-a91e-438b-832b-b9ba15c87e15-kube-api-access-mk4zs\") pod \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.585626 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.585678 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-config-data\") pod \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.585696 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-scripts\") pod \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.585743 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-httpd-run\") pod \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.585794 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-logs\") pod \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\" (UID: \"e48a6f8b-a91e-438b-832b-b9ba15c87e15\") " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.587877 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e48a6f8b-a91e-438b-832b-b9ba15c87e15" (UID: "e48a6f8b-a91e-438b-832b-b9ba15c87e15"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.588128 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-logs" (OuterVolumeSpecName: "logs") pod "e48a6f8b-a91e-438b-832b-b9ba15c87e15" (UID: "e48a6f8b-a91e-438b-832b-b9ba15c87e15"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.594589 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.601204 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.191724605 podStartE2EDuration="20.601178964s" podCreationTimestamp="2026-02-15 20:36:11 +0000 UTC" firstStartedPulling="2026-02-15 20:36:12.2636547 +0000 UTC m=+1180.129670323" lastFinishedPulling="2026-02-15 20:36:29.673109059 +0000 UTC m=+1197.539124682" observedRunningTime="2026-02-15 20:36:31.5512117 +0000 UTC m=+1199.417227323" watchObservedRunningTime="2026-02-15 20:36:31.601178964 +0000 UTC m=+1199.467194577" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.615035 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e48a6f8b-a91e-438b-832b-b9ba15c87e15","Type":"ContainerDied","Data":"cad371487c069f9da3937fb49038d5395b203f526bda6393332258698c6ee339"} Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.615229 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.623145 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b7b86987-zpgc8" event={"ID":"843b8077-b23a-42f9-bf70-79767ce35b4f","Type":"ContainerStarted","Data":"ae67eaf7d42802513de3a2ae3715fd19a9cdd42fe6a2ff6bb94d199db371fb74"} Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629089 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629446 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerName="glance-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629463 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerName="glance-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629480 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e551f68-1918-4b4a-99de-c7e05c507187" containerName="neutron-api" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629486 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e551f68-1918-4b4a-99de-c7e05c507187" containerName="neutron-api" Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629502 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="sg-core" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629508 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="sg-core" Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629517 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e551f68-1918-4b4a-99de-c7e05c507187" containerName="neutron-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629523 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e551f68-1918-4b4a-99de-c7e05c507187" containerName="neutron-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629536 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="ceilometer-central-agent" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629541 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="ceilometer-central-agent" Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629549 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerName="glance-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629554 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerName="glance-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629565 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="proxy-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629571 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="proxy-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629583 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerName="glance-log" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629589 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerName="glance-log" Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629598 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerName="glance-log" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629603 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerName="glance-log" Feb 15 20:36:31 crc kubenswrapper[4735]: E0215 20:36:31.629614 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="ceilometer-notification-agent" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629620 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="ceilometer-notification-agent" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629773 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerName="glance-log" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629784 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e551f68-1918-4b4a-99de-c7e05c507187" containerName="neutron-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629797 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerName="glance-log" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629807 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="sg-core" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629818 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" containerName="glance-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629827 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e551f68-1918-4b4a-99de-c7e05c507187" containerName="neutron-api" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629837 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" containerName="glance-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629849 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="ceilometer-central-agent" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629856 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="proxy-httpd" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.629865 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" containerName="ceilometer-notification-agent" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.630791 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.640961 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "e48a6f8b-a91e-438b-832b-b9ba15c87e15" (UID: "e48a6f8b-a91e-438b-832b-b9ba15c87e15"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.644273 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.644526 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.646833 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-scripts" (OuterVolumeSpecName: "scripts") pod "e48a6f8b-a91e-438b-832b-b9ba15c87e15" (UID: "e48a6f8b-a91e-438b-832b-b9ba15c87e15"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.648559 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.652186 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e48a6f8b-a91e-438b-832b-b9ba15c87e15-kube-api-access-mk4zs" (OuterVolumeSpecName: "kube-api-access-mk4zs") pod "e48a6f8b-a91e-438b-832b-b9ba15c87e15" (UID: "e48a6f8b-a91e-438b-832b-b9ba15c87e15"). InnerVolumeSpecName "kube-api-access-mk4zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.688300 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.688394 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.688544 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v227b\" (UniqueName: \"kubernetes.io/projected/91998299-81cc-409a-a32d-2d2fa671d379-kube-api-access-v227b\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.688633 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.688879 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.688997 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91998299-81cc-409a-a32d-2d2fa671d379-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.689229 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.689306 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91998299-81cc-409a-a32d-2d2fa671d379-logs\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.689380 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.689398 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e48a6f8b-a91e-438b-832b-b9ba15c87e15-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.689407 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk4zs\" (UniqueName: \"kubernetes.io/projected/e48a6f8b-a91e-438b-832b-b9ba15c87e15-kube-api-access-mk4zs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.689449 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.689460 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.763545 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e48a6f8b-a91e-438b-832b-b9ba15c87e15" (UID: "e48a6f8b-a91e-438b-832b-b9ba15c87e15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.767294 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.768261 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-config-data" (OuterVolumeSpecName: "config-data") pod "e48a6f8b-a91e-438b-832b-b9ba15c87e15" (UID: "e48a6f8b-a91e-438b-832b-b9ba15c87e15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.791546 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91998299-81cc-409a-a32d-2d2fa671d379-logs\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.791583 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.791612 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.791660 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v227b\" (UniqueName: \"kubernetes.io/projected/91998299-81cc-409a-a32d-2d2fa671d379-kube-api-access-v227b\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.791684 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.791716 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.791767 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91998299-81cc-409a-a32d-2d2fa671d379-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.791815 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.792633 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.793937 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91998299-81cc-409a-a32d-2d2fa671d379-logs\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.793997 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.798620 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91998299-81cc-409a-a32d-2d2fa671d379-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.799134 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.802319 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.803880 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.814632 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.821116 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v227b\" (UniqueName: \"kubernetes.io/projected/91998299-81cc-409a-a32d-2d2fa671d379-kube-api-access-v227b\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.822517 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.823617 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91998299-81cc-409a-a32d-2d2fa671d379-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.841160 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-f66hh"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.849265 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e48a6f8b-a91e-438b-832b-b9ba15c87e15" (UID: "e48a6f8b-a91e-438b-832b-b9ba15c87e15"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.855956 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"91998299-81cc-409a-a32d-2d2fa671d379\") " pod="openstack/glance-default-internal-api-0" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.876676 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-m6ztx"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.904705 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e48a6f8b-a91e-438b-832b-b9ba15c87e15-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.933214 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-33f9-account-create-update-t72d7"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.945391 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d824-account-create-update-gmj74"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.963904 4735 scope.go:117] "RemoveContainer" containerID="51c478a3412d9dd59ac3624eb9e78d9ded5143fcdf90a0f76ed2fc9271a76bf1" Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.976997 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-af86-account-create-update-5mp2z"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.987925 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-59f96fd876-q78bj"] Feb 15 20:36:31 crc kubenswrapper[4735]: I0215 20:36:31.998425 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-59f96fd876-q78bj"] Feb 15 20:36:32 crc kubenswrapper[4735]: E0215 20:36:32.058167 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a00bb1_fbaf_45a7_882b_6a2a24bd100d.slice/crio-conmon-8d114b6173c49bd9cb2de87ba812abf6cde86c1ef7d83bcbe0aaff2a24425584.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e551f68_1918_4b4a_99de_c7e05c507187.slice/crio-7c06fc1659d45c78cd1d29f4b44200f8c4675dc45011376308cfbc0b5abfbba7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bdadc81_14d1_469b_82c1_68f79cfa71a9.slice/crio-0eceef46b722125a7a67a2e0b988daa2f542ecc7fd172b767ea9348df61f5f01.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a00bb1_fbaf_45a7_882b_6a2a24bd100d.slice/crio-73e5563131beea3e291af24d55b00d086bcd2089890020e6c943154f24febfce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod107aa5db_f0d0_4817_8d2d_12da89492473.slice/crio-conmon-be1dc831e7060f2c2a416d0785c4bfb76e59d77ebfaab5dd8be8cff4c319dd71.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43de4639_a922_4182_9377_d1f28b3aa4c1.slice/crio-conmon-94228183c3d1abcbd5b788cf2f92ec757e11f8d18a65dfdd44912dcfc7991e88.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode48a6f8b_a91e_438b_832b_b9ba15c87e15.slice/crio-conmon-7f220f4630db2ac2af4c43eee7e69a114fc9d4b048e5ff25b90aaa27e470205d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bdadc81_14d1_469b_82c1_68f79cfa71a9.slice/crio-2362903ea75605e58912d9ab9e9910e689415b6e37485ede12f5837307f8d68a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e551f68_1918_4b4a_99de_c7e05c507187.slice/crio-conmon-51c478a3412d9dd59ac3624eb9e78d9ded5143fcdf90a0f76ed2fc9271a76bf1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a00bb1_fbaf_45a7_882b_6a2a24bd100d.slice/crio-conmon-6c226ff20ed0038cef94ee50f1214e3a0e63d9739012d1195ec9cb4b02b520c1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod970d5293_b621_45e1_a1e4_8cc176c9a148.slice/crio-conmon-063addf10589b9afd3c588e4171de6d504c19a3cbf44f266f86a25dfe5ca99e3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f788fda_f0ae_45f7_a233_8e4963e5ed2d.slice/crio-conmon-509c549dfe02bd37ca623752f3760afb9e5b4c41da36b238f99fb78d66aac256.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bdadc81_14d1_469b_82c1_68f79cfa71a9.slice/crio-conmon-0eceef46b722125a7a67a2e0b988daa2f542ecc7fd172b767ea9348df61f5f01.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a00bb1_fbaf_45a7_882b_6a2a24bd100d.slice/crio-6c226ff20ed0038cef94ee50f1214e3a0e63d9739012d1195ec9cb4b02b520c1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a00bb1_fbaf_45a7_882b_6a2a24bd100d.slice/crio-563cd8683e20b062dc2adde56b23ee47afd98fa9e99b53d61e87b491b24d6835.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43de4639_a922_4182_9377_d1f28b3aa4c1.slice/crio-94228183c3d1abcbd5b788cf2f92ec757e11f8d18a65dfdd44912dcfc7991e88.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode48a6f8b_a91e_438b_832b_b9ba15c87e15.slice/crio-fce032895e4ab9a42d9c005265ba71abf97dc38fd3a315afabcc1cd553214a3e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bdadc81_14d1_469b_82c1_68f79cfa71a9.slice/crio-conmon-2362903ea75605e58912d9ab9e9910e689415b6e37485ede12f5837307f8d68a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod970d5293_b621_45e1_a1e4_8cc176c9a148.slice/crio-063addf10589b9afd3c588e4171de6d504c19a3cbf44f266f86a25dfe5ca99e3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a00bb1_fbaf_45a7_882b_6a2a24bd100d.slice/crio-8d114b6173c49bd9cb2de87ba812abf6cde86c1ef7d83bcbe0aaff2a24425584.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode48a6f8b_a91e_438b_832b_b9ba15c87e15.slice/crio-conmon-fce032895e4ab9a42d9c005265ba71abf97dc38fd3a315afabcc1cd553214a3e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod107aa5db_f0d0_4817_8d2d_12da89492473.slice/crio-be1dc831e7060f2c2a416d0785c4bfb76e59d77ebfaab5dd8be8cff4c319dd71.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76a8192d_7bcd_4f4a_b798_907d67b3d21c.slice/crio-02617330b3775ed02a8aa95ad89b3c2f5d4d45f1e9a94e6337caf90f424865fd\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode48a6f8b_a91e_438b_832b_b9ba15c87e15.slice/crio-7f220f4630db2ac2af4c43eee7e69a114fc9d4b048e5ff25b90aaa27e470205d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e551f68_1918_4b4a_99de_c7e05c507187.slice/crio-conmon-7c06fc1659d45c78cd1d29f4b44200f8c4675dc45011376308cfbc0b5abfbba7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a00bb1_fbaf_45a7_882b_6a2a24bd100d.slice/crio-conmon-563cd8683e20b062dc2adde56b23ee47afd98fa9e99b53d61e87b491b24d6835.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a00bb1_fbaf_45a7_882b_6a2a24bd100d.slice/crio-conmon-73e5563131beea3e291af24d55b00d086bcd2089890020e6c943154f24febfce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f788fda_f0ae_45f7_a233_8e4963e5ed2d.slice/crio-509c549dfe02bd37ca623752f3760afb9e5b4c41da36b238f99fb78d66aac256.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bdadc81_14d1_469b_82c1_68f79cfa71a9.slice\": RecentStats: unable to find data in memory cache]" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.087837 4735 scope.go:117] "RemoveContainer" containerID="7c06fc1659d45c78cd1d29f4b44200f8c4675dc45011376308cfbc0b5abfbba7" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.138775 4735 scope.go:117] "RemoveContainer" containerID="73e5563131beea3e291af24d55b00d086bcd2089890020e6c943154f24febfce" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.186594 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.224230 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data\") pod \"107aa5db-f0d0-4817-8d2d-12da89492473\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.224298 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data-custom\") pod \"107aa5db-f0d0-4817-8d2d-12da89492473\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.224405 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-scripts\") pod \"107aa5db-f0d0-4817-8d2d-12da89492473\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.224497 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/107aa5db-f0d0-4817-8d2d-12da89492473-etc-machine-id\") pod \"107aa5db-f0d0-4817-8d2d-12da89492473\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.224559 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-combined-ca-bundle\") pod \"107aa5db-f0d0-4817-8d2d-12da89492473\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.224590 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/107aa5db-f0d0-4817-8d2d-12da89492473-logs\") pod \"107aa5db-f0d0-4817-8d2d-12da89492473\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.224645 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k7gb\" (UniqueName: \"kubernetes.io/projected/107aa5db-f0d0-4817-8d2d-12da89492473-kube-api-access-9k7gb\") pod \"107aa5db-f0d0-4817-8d2d-12da89492473\" (UID: \"107aa5db-f0d0-4817-8d2d-12da89492473\") " Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.227413 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/107aa5db-f0d0-4817-8d2d-12da89492473-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "107aa5db-f0d0-4817-8d2d-12da89492473" (UID: "107aa5db-f0d0-4817-8d2d-12da89492473"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.227828 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/107aa5db-f0d0-4817-8d2d-12da89492473-logs" (OuterVolumeSpecName: "logs") pod "107aa5db-f0d0-4817-8d2d-12da89492473" (UID: "107aa5db-f0d0-4817-8d2d-12da89492473"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.241893 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.257341 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-scripts" (OuterVolumeSpecName: "scripts") pod "107aa5db-f0d0-4817-8d2d-12da89492473" (UID: "107aa5db-f0d0-4817-8d2d-12da89492473"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.290080 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.302894 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "107aa5db-f0d0-4817-8d2d-12da89492473" (UID: "107aa5db-f0d0-4817-8d2d-12da89492473"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.330503 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/107aa5db-f0d0-4817-8d2d-12da89492473-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.330667 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.330676 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.330684 4735 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/107aa5db-f0d0-4817-8d2d-12da89492473-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.339849 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.343605 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/107aa5db-f0d0-4817-8d2d-12da89492473-kube-api-access-9k7gb" (OuterVolumeSpecName: "kube-api-access-9k7gb") pod "107aa5db-f0d0-4817-8d2d-12da89492473" (UID: "107aa5db-f0d0-4817-8d2d-12da89492473"). InnerVolumeSpecName "kube-api-access-9k7gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.363623 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.376603 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:36:32 crc kubenswrapper[4735]: E0215 20:36:32.385267 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="107aa5db-f0d0-4817-8d2d-12da89492473" containerName="cinder-api-log" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.385306 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="107aa5db-f0d0-4817-8d2d-12da89492473" containerName="cinder-api-log" Feb 15 20:36:32 crc kubenswrapper[4735]: E0215 20:36:32.385348 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="107aa5db-f0d0-4817-8d2d-12da89492473" containerName="cinder-api" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.385354 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="107aa5db-f0d0-4817-8d2d-12da89492473" containerName="cinder-api" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.385627 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="107aa5db-f0d0-4817-8d2d-12da89492473" containerName="cinder-api" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.385662 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="107aa5db-f0d0-4817-8d2d-12da89492473" containerName="cinder-api-log" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.386579 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.389480 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.390387 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.395054 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.398195 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.405811 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.410651 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.423477 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432283 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-config-data\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432330 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a32ed00-6692-40df-af48-ecd34af2ffd2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432355 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-scripts\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432387 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxshk\" (UniqueName: \"kubernetes.io/projected/0bcd5cef-2249-4a58-ac25-db85ed772900-kube-api-access-vxshk\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432504 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432551 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432589 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-scripts\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432693 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-config-data\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432719 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a32ed00-6692-40df-af48-ecd34af2ffd2-logs\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432748 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432777 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-run-httpd\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432821 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r5fr\" (UniqueName: \"kubernetes.io/projected/3a32ed00-6692-40df-af48-ecd34af2ffd2-kube-api-access-6r5fr\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432840 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.432934 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.433010 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-log-httpd\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.433302 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k7gb\" (UniqueName: \"kubernetes.io/projected/107aa5db-f0d0-4817-8d2d-12da89492473-kube-api-access-9k7gb\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.436670 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.490327 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "107aa5db-f0d0-4817-8d2d-12da89492473" (UID: "107aa5db-f0d0-4817-8d2d-12da89492473"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.546831 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a32ed00-6692-40df-af48-ecd34af2ffd2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.547058 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-scripts\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.547503 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a32ed00-6692-40df-af48-ecd34af2ffd2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.547917 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxshk\" (UniqueName: \"kubernetes.io/projected/0bcd5cef-2249-4a58-ac25-db85ed772900-kube-api-access-vxshk\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.561133 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.561617 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.561705 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-scripts\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.561880 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-config-data\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.562079 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a32ed00-6692-40df-af48-ecd34af2ffd2-logs\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.562412 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.562504 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-run-httpd\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.562591 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r5fr\" (UniqueName: \"kubernetes.io/projected/3a32ed00-6692-40df-af48-ecd34af2ffd2-kube-api-access-6r5fr\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.562667 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.562778 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.562881 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-log-httpd\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.563038 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-config-data\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.563194 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.564600 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.566694 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-scripts\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.577378 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-scripts\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.578058 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.580594 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-config-data\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.581495 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-run-httpd\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.582211 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-log-httpd\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.582614 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a32ed00-6692-40df-af48-ecd34af2ffd2-logs\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.591597 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.627491 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.629639 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.630628 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r5fr\" (UniqueName: \"kubernetes.io/projected/3a32ed00-6692-40df-af48-ecd34af2ffd2-kube-api-access-6r5fr\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.634787 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxshk\" (UniqueName: \"kubernetes.io/projected/0bcd5cef-2249-4a58-ac25-db85ed772900-kube-api-access-vxshk\") pod \"ceilometer-0\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " pod="openstack/ceilometer-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.662215 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a32ed00-6692-40df-af48-ecd34af2ffd2-config-data\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.670151 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f66hh" event={"ID":"85484764-8eab-4bbd-8d2f-1726f287ada3","Type":"ContainerStarted","Data":"a165350886231990f6f8a1900202ef0c5ef107626c77012da09ab5c9a593d60f"} Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.688867 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-m6ztx" event={"ID":"48189563-f3c7-49a0-b01e-86b25b077fd5","Type":"ContainerStarted","Data":"0f1548d194a99f08acec43643a3412f38380dea230ed3068fa946c56108d1ea7"} Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.689097 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-m6ztx" event={"ID":"48189563-f3c7-49a0-b01e-86b25b077fd5","Type":"ContainerStarted","Data":"d651b1dd9136ac75ae4c8a415ce1088cba54470e8690f04c20d62c8446fff5ae"} Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.694176 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data" (OuterVolumeSpecName: "config-data") pod "107aa5db-f0d0-4817-8d2d-12da89492473" (UID: "107aa5db-f0d0-4817-8d2d-12da89492473"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.716263 4735 generic.go:334] "Generic (PLEG): container finished" podID="c8507ec7-d881-4a4a-82dc-3af56d4d53aa" containerID="3bc53ecea541aa831649411facd7ba4a1417813f89093f199f69eae98e61b2f8" exitCode=0 Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.716331 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-svbdx" event={"ID":"c8507ec7-d881-4a4a-82dc-3af56d4d53aa","Type":"ContainerDied","Data":"3bc53ecea541aa831649411facd7ba4a1417813f89093f199f69eae98e61b2f8"} Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.747594 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3a32ed00-6692-40df-af48-ecd34af2ffd2\") " pod="openstack/glance-default-external-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.748053 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-33f9-account-create-update-t72d7" event={"ID":"493558c5-9375-43a2-92e1-9f20470fc681","Type":"ContainerStarted","Data":"d05b84182198b7b8e1f0986b7c275d398ccf3f0a0ca6b535a36c61d8726a7ae9"} Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.752847 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-m6ztx" podStartSLOduration=3.752824601 podStartE2EDuration="3.752824601s" podCreationTimestamp="2026-02-15 20:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:32.740239661 +0000 UTC m=+1200.606255284" watchObservedRunningTime="2026-02-15 20:36:32.752824601 +0000 UTC m=+1200.618840224" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.767367 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/107aa5db-f0d0-4817-8d2d-12da89492473-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.810449 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b7b86987-zpgc8" event={"ID":"843b8077-b23a-42f9-bf70-79767ce35b4f","Type":"ContainerStarted","Data":"535813ed2776054bb9913bbad57e620e59558db3c570b52c37ba960640588ef9"} Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.834363 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"107aa5db-f0d0-4817-8d2d-12da89492473","Type":"ContainerDied","Data":"5872ffbc7b59e977da9d19ee421270cdc341d454bac69a3db14c9d0c7ce65559"} Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.834464 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.850205 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.913206 4735 scope.go:117] "RemoveContainer" containerID="6c226ff20ed0038cef94ee50f1214e3a0e63d9739012d1195ec9cb4b02b520c1" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.925884 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bdadc81-14d1-469b-82c1-68f79cfa71a9" path="/var/lib/kubelet/pods/1bdadc81-14d1-469b-82c1-68f79cfa71a9/volumes" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.926699 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49a00bb1-fbaf-45a7-882b-6a2a24bd100d" path="/var/lib/kubelet/pods/49a00bb1-fbaf-45a7-882b-6a2a24bd100d/volumes" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.942913 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e551f68-1918-4b4a-99de-c7e05c507187" path="/var/lib/kubelet/pods/4e551f68-1918-4b4a-99de-c7e05c507187/volumes" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.943616 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e48a6f8b-a91e-438b-832b-b9ba15c87e15" path="/var/lib/kubelet/pods/e48a6f8b-a91e-438b-832b-b9ba15c87e15/volumes" Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.963359 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-af86-account-create-update-5mp2z" event={"ID":"f27472c1-f67e-49e6-bd56-a36101f69c64","Type":"ContainerStarted","Data":"3575f358d1444f740aaa95f505bad82de163e768dea69ae76f98254a5e6b74bf"} Feb 15 20:36:32 crc kubenswrapper[4735]: I0215 20:36:32.963413 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d824-account-create-update-gmj74" event={"ID":"f2d80b54-e0f6-415a-9bab-1cff8e208efa","Type":"ContainerStarted","Data":"7222645eea1f465a186e03ef416b71868ff9eed53957eac51dcc4ba3c0f99ea7"} Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.026781 4735 scope.go:117] "RemoveContainer" containerID="563cd8683e20b062dc2adde56b23ee47afd98fa9e99b53d61e87b491b24d6835" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.136689 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.177398 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.256456 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.260008 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.274298 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.274442 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.274554 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.285929 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f3720a1-390e-457d-a4c2-d758b04a90d3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.286006 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.286042 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.286061 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-config-data-custom\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.286077 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-config-data\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.286094 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.286109 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-scripts\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.286129 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf9df\" (UniqueName: \"kubernetes.io/projected/1f3720a1-390e-457d-a4c2-d758b04a90d3-kube-api-access-rf9df\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.286226 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3720a1-390e-457d-a4c2-d758b04a90d3-logs\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.314530 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.387494 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-scripts\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.387531 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf9df\" (UniqueName: \"kubernetes.io/projected/1f3720a1-390e-457d-a4c2-d758b04a90d3-kube-api-access-rf9df\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.387633 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3720a1-390e-457d-a4c2-d758b04a90d3-logs\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.387650 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f3720a1-390e-457d-a4c2-d758b04a90d3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.387682 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.387711 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.387728 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-config-data-custom\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.387746 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-config-data\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.387763 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.389063 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f3720a1-390e-457d-a4c2-d758b04a90d3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.394871 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3720a1-390e-457d-a4c2-d758b04a90d3-logs\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.399153 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-scripts\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.415027 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-config-data-custom\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.415122 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.415501 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.416053 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.416572 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3720a1-390e-457d-a4c2-d758b04a90d3-config-data\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.431558 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf9df\" (UniqueName: \"kubernetes.io/projected/1f3720a1-390e-457d-a4c2-d758b04a90d3-kube-api-access-rf9df\") pod \"cinder-api-0\" (UID: \"1f3720a1-390e-457d-a4c2-d758b04a90d3\") " pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.558001 4735 scope.go:117] "RemoveContainer" containerID="8d114b6173c49bd9cb2de87ba812abf6cde86c1ef7d83bcbe0aaff2a24425584" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.627862 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.646166 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.664462 4735 scope.go:117] "RemoveContainer" containerID="fce032895e4ab9a42d9c005265ba71abf97dc38fd3a315afabcc1cd553214a3e" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.677856 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.900101 4735 scope.go:117] "RemoveContainer" containerID="7f220f4630db2ac2af4c43eee7e69a114fc9d4b048e5ff25b90aaa27e470205d" Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.938132 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-33f9-account-create-update-t72d7" event={"ID":"493558c5-9375-43a2-92e1-9f20470fc681","Type":"ContainerStarted","Data":"8a23895043faca9a454ecf4694ebaee4530963d9005db42b7838e1641d8c4a1c"} Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.979327 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 15 20:36:33 crc kubenswrapper[4735]: I0215 20:36:33.989025 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-33f9-account-create-update-t72d7" podStartSLOduration=4.98900321 podStartE2EDuration="4.98900321s" podCreationTimestamp="2026-02-15 20:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:33.961393622 +0000 UTC m=+1201.827409245" watchObservedRunningTime="2026-02-15 20:36:33.98900321 +0000 UTC m=+1201.855018843" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.008253 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.008292 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:34 crc kubenswrapper[4735]: W0215 20:36:34.015252 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91998299_81cc_409a_a32d_2d2fa671d379.slice/crio-7bca75b5e1da0153a7110f944eea9bb5219b3a6c1b16663a7f3f901d426a4ef7 WatchSource:0}: Error finding container 7bca75b5e1da0153a7110f944eea9bb5219b3a6c1b16663a7f3f901d426a4ef7: Status 404 returned error can't find the container with id 7bca75b5e1da0153a7110f944eea9bb5219b3a6c1b16663a7f3f901d426a4ef7 Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.027779 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-af86-account-create-update-5mp2z" event={"ID":"f27472c1-f67e-49e6-bd56-a36101f69c64","Type":"ContainerStarted","Data":"c30cb3b3720eae317ee91f199f741bd74fe72b9894661fd98f2c3bbb66e16e83"} Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.037013 4735 scope.go:117] "RemoveContainer" containerID="be1dc831e7060f2c2a416d0785c4bfb76e59d77ebfaab5dd8be8cff4c319dd71" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.057365 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-b7b86987-zpgc8" podStartSLOduration=14.05734693 podStartE2EDuration="14.05734693s" podCreationTimestamp="2026-02-15 20:36:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:34.044119501 +0000 UTC m=+1201.910135124" watchObservedRunningTime="2026-02-15 20:36:34.05734693 +0000 UTC m=+1201.923362543" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.067106 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f66hh" event={"ID":"85484764-8eab-4bbd-8d2f-1726f287ada3","Type":"ContainerStarted","Data":"26b9cc975a1348b164a681f96d9d5e400fbc845f81921d89ce8555b9b5dcba7a"} Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.070540 4735 generic.go:334] "Generic (PLEG): container finished" podID="48189563-f3c7-49a0-b01e-86b25b077fd5" containerID="0f1548d194a99f08acec43643a3412f38380dea230ed3068fa946c56108d1ea7" exitCode=0 Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.070753 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-m6ztx" event={"ID":"48189563-f3c7-49a0-b01e-86b25b077fd5","Type":"ContainerDied","Data":"0f1548d194a99f08acec43643a3412f38380dea230ed3068fa946c56108d1ea7"} Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.090904 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-af86-account-create-update-5mp2z" podStartSLOduration=5.090887148 podStartE2EDuration="5.090887148s" podCreationTimestamp="2026-02-15 20:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:34.06434015 +0000 UTC m=+1201.930355773" watchObservedRunningTime="2026-02-15 20:36:34.090887148 +0000 UTC m=+1201.956902761" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.162097 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-d824-account-create-update-gmj74" podStartSLOduration=5.162078816 podStartE2EDuration="5.162078816s" podCreationTimestamp="2026-02-15 20:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:34.082709827 +0000 UTC m=+1201.948725450" watchObservedRunningTime="2026-02-15 20:36:34.162078816 +0000 UTC m=+1202.028094439" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.200996 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-f66hh" podStartSLOduration=5.200975509 podStartE2EDuration="5.200975509s" podCreationTimestamp="2026-02-15 20:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:34.125358322 +0000 UTC m=+1201.991373965" watchObservedRunningTime="2026-02-15 20:36:34.200975509 +0000 UTC m=+1202.066991132" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.243279 4735 scope.go:117] "RemoveContainer" containerID="752d9da1cf37173f495550234a7f8e573937d3f4e6f9151ac82a2829ac2cd97f" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.629497 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.779067 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.779983 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.784063 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.851430 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.851908 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:36:34 crc kubenswrapper[4735]: I0215 20:36:34.937566 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="107aa5db-f0d0-4817-8d2d-12da89492473" path="/var/lib/kubelet/pods/107aa5db-f0d0-4817-8d2d-12da89492473/volumes" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.003614 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.034202 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.164327 4735 generic.go:334] "Generic (PLEG): container finished" podID="f27472c1-f67e-49e6-bd56-a36101f69c64" containerID="c30cb3b3720eae317ee91f199f741bd74fe72b9894661fd98f2c3bbb66e16e83" exitCode=0 Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.164812 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-af86-account-create-update-5mp2z" event={"ID":"f27472c1-f67e-49e6-bd56-a36101f69c64","Type":"ContainerDied","Data":"c30cb3b3720eae317ee91f199f741bd74fe72b9894661fd98f2c3bbb66e16e83"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.169427 4735 generic.go:334] "Generic (PLEG): container finished" podID="85484764-8eab-4bbd-8d2f-1726f287ada3" containerID="26b9cc975a1348b164a681f96d9d5e400fbc845f81921d89ce8555b9b5dcba7a" exitCode=0 Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.169483 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f66hh" event={"ID":"85484764-8eab-4bbd-8d2f-1726f287ada3","Type":"ContainerDied","Data":"26b9cc975a1348b164a681f96d9d5e400fbc845f81921d89ce8555b9b5dcba7a"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.170144 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-operator-scripts\") pod \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\" (UID: \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\") " Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.170206 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvz4w\" (UniqueName: \"kubernetes.io/projected/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-kube-api-access-mvz4w\") pod \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\" (UID: \"c8507ec7-d881-4a4a-82dc-3af56d4d53aa\") " Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.173798 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c8507ec7-d881-4a4a-82dc-3af56d4d53aa" (UID: "c8507ec7-d881-4a4a-82dc-3af56d4d53aa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.183648 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-kube-api-access-mvz4w" (OuterVolumeSpecName: "kube-api-access-mvz4w") pod "c8507ec7-d881-4a4a-82dc-3af56d4d53aa" (UID: "c8507ec7-d881-4a4a-82dc-3af56d4d53aa"). InnerVolumeSpecName "kube-api-access-mvz4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.188976 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-svbdx" event={"ID":"c8507ec7-d881-4a4a-82dc-3af56d4d53aa","Type":"ContainerDied","Data":"47839f57575740c82944bf7680f988009debb423027efe1dfb3d6c662e104530"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.189010 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47839f57575740c82944bf7680f988009debb423027efe1dfb3d6c662e104530" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.189069 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-svbdx" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.219153 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1f3720a1-390e-457d-a4c2-d758b04a90d3","Type":"ContainerStarted","Data":"9a8c8c0d251beff2fa26c633226387d166bf6c9f6d8fa6d83744c0c051b0ee1e"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.241589 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b7b86987-zpgc8" event={"ID":"843b8077-b23a-42f9-bf70-79767ce35b4f","Type":"ContainerStarted","Data":"be310fe6b79ee7e738a6fd316c8e676b4eeb38c587acb0f5132896f3710d81ad"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.252230 4735 generic.go:334] "Generic (PLEG): container finished" podID="f2d80b54-e0f6-415a-9bab-1cff8e208efa" containerID="83c1e4a202bd81bfb86878c03dfdc78c9f43f793413d3a965a726ae68500d2a6" exitCode=0 Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.252300 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d824-account-create-update-gmj74" event={"ID":"f2d80b54-e0f6-415a-9bab-1cff8e208efa","Type":"ContainerDied","Data":"83c1e4a202bd81bfb86878c03dfdc78c9f43f793413d3a965a726ae68500d2a6"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.262712 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerStarted","Data":"ae59437b6f6cb0d10a42840eb070de10e541f029ef61d68a312b5c32626f992f"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.274137 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.274169 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvz4w\" (UniqueName: \"kubernetes.io/projected/c8507ec7-d881-4a4a-82dc-3af56d4d53aa-kube-api-access-mvz4w\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.274483 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3a32ed00-6692-40df-af48-ecd34af2ffd2","Type":"ContainerStarted","Data":"6a3a3623bdd4bf1b9e916abb1988f00027888a46e4c1493ea471fd55a6f37cce"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.297390 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91998299-81cc-409a-a32d-2d2fa671d379","Type":"ContainerStarted","Data":"7bca75b5e1da0153a7110f944eea9bb5219b3a6c1b16663a7f3f901d426a4ef7"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.300833 4735 generic.go:334] "Generic (PLEG): container finished" podID="493558c5-9375-43a2-92e1-9f20470fc681" containerID="8a23895043faca9a454ecf4694ebaee4530963d9005db42b7838e1641d8c4a1c" exitCode=0 Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.300970 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-33f9-account-create-update-t72d7" event={"ID":"493558c5-9375-43a2-92e1-9f20470fc681","Type":"ContainerDied","Data":"8a23895043faca9a454ecf4694ebaee4530963d9005db42b7838e1641d8c4a1c"} Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.717271 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.783597 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v99f\" (UniqueName: \"kubernetes.io/projected/48189563-f3c7-49a0-b01e-86b25b077fd5-kube-api-access-8v99f\") pod \"48189563-f3c7-49a0-b01e-86b25b077fd5\" (UID: \"48189563-f3c7-49a0-b01e-86b25b077fd5\") " Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.783757 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48189563-f3c7-49a0-b01e-86b25b077fd5-operator-scripts\") pod \"48189563-f3c7-49a0-b01e-86b25b077fd5\" (UID: \"48189563-f3c7-49a0-b01e-86b25b077fd5\") " Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.784259 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48189563-f3c7-49a0-b01e-86b25b077fd5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "48189563-f3c7-49a0-b01e-86b25b077fd5" (UID: "48189563-f3c7-49a0-b01e-86b25b077fd5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.806369 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48189563-f3c7-49a0-b01e-86b25b077fd5-kube-api-access-8v99f" (OuterVolumeSpecName: "kube-api-access-8v99f") pod "48189563-f3c7-49a0-b01e-86b25b077fd5" (UID: "48189563-f3c7-49a0-b01e-86b25b077fd5"). InnerVolumeSpecName "kube-api-access-8v99f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.886070 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v99f\" (UniqueName: \"kubernetes.io/projected/48189563-f3c7-49a0-b01e-86b25b077fd5-kube-api-access-8v99f\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:35 crc kubenswrapper[4735]: I0215 20:36:35.886456 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48189563-f3c7-49a0-b01e-86b25b077fd5-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:36 crc kubenswrapper[4735]: I0215 20:36:36.363734 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-m6ztx" event={"ID":"48189563-f3c7-49a0-b01e-86b25b077fd5","Type":"ContainerDied","Data":"d651b1dd9136ac75ae4c8a415ce1088cba54470e8690f04c20d62c8446fff5ae"} Feb 15 20:36:36 crc kubenswrapper[4735]: I0215 20:36:36.363996 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d651b1dd9136ac75ae4c8a415ce1088cba54470e8690f04c20d62c8446fff5ae" Feb 15 20:36:36 crc kubenswrapper[4735]: I0215 20:36:36.364047 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-m6ztx" Feb 15 20:36:36 crc kubenswrapper[4735]: I0215 20:36:36.405390 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerStarted","Data":"92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c"} Feb 15 20:36:36 crc kubenswrapper[4735]: I0215 20:36:36.505261 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91998299-81cc-409a-a32d-2d2fa671d379","Type":"ContainerStarted","Data":"b801fbac713d0fd676ab73ba3b2739c8dbca561ff75e3a459d3fb98729ba680e"} Feb 15 20:36:36 crc kubenswrapper[4735]: I0215 20:36:36.528055 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3a32ed00-6692-40df-af48-ecd34af2ffd2","Type":"ContainerStarted","Data":"9dce0ddf46f427f2af7020739ee2e419be2a14437fa27414c6f1013551a5f5b7"} Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.240572 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.433700 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29x5h\" (UniqueName: \"kubernetes.io/projected/f27472c1-f67e-49e6-bd56-a36101f69c64-kube-api-access-29x5h\") pod \"f27472c1-f67e-49e6-bd56-a36101f69c64\" (UID: \"f27472c1-f67e-49e6-bd56-a36101f69c64\") " Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.433881 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27472c1-f67e-49e6-bd56-a36101f69c64-operator-scripts\") pod \"f27472c1-f67e-49e6-bd56-a36101f69c64\" (UID: \"f27472c1-f67e-49e6-bd56-a36101f69c64\") " Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.434804 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27472c1-f67e-49e6-bd56-a36101f69c64-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f27472c1-f67e-49e6-bd56-a36101f69c64" (UID: "f27472c1-f67e-49e6-bd56-a36101f69c64"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.470199 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27472c1-f67e-49e6-bd56-a36101f69c64-kube-api-access-29x5h" (OuterVolumeSpecName: "kube-api-access-29x5h") pod "f27472c1-f67e-49e6-bd56-a36101f69c64" (UID: "f27472c1-f67e-49e6-bd56-a36101f69c64"). InnerVolumeSpecName "kube-api-access-29x5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.542329 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27472c1-f67e-49e6-bd56-a36101f69c64-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.542373 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29x5h\" (UniqueName: \"kubernetes.io/projected/f27472c1-f67e-49e6-bd56-a36101f69c64-kube-api-access-29x5h\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.593086 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-af86-account-create-update-5mp2z" event={"ID":"f27472c1-f67e-49e6-bd56-a36101f69c64","Type":"ContainerDied","Data":"3575f358d1444f740aaa95f505bad82de163e768dea69ae76f98254a5e6b74bf"} Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.593371 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3575f358d1444f740aaa95f505bad82de163e768dea69ae76f98254a5e6b74bf" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.593442 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-af86-account-create-update-5mp2z" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.634863 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerStarted","Data":"8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00"} Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.666634 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1f3720a1-390e-457d-a4c2-d758b04a90d3","Type":"ContainerStarted","Data":"f311b59a308073c7104950ce14e7c9f16bbc2e1e0239a6e1ab5bf90aa337f909"} Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.682366 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.686798 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-33f9-account-create-update-t72d7" event={"ID":"493558c5-9375-43a2-92e1-9f20470fc681","Type":"ContainerDied","Data":"d05b84182198b7b8e1f0986b7c275d398ccf3f0a0ca6b535a36c61d8726a7ae9"} Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.686839 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d05b84182198b7b8e1f0986b7c275d398ccf3f0a0ca6b535a36c61d8726a7ae9" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.694129 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.707599 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.728314 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.728290593 podStartE2EDuration="6.728290593s" podCreationTimestamp="2026-02-15 20:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:37.703158353 +0000 UTC m=+1205.569173976" watchObservedRunningTime="2026-02-15 20:36:37.728290593 +0000 UTC m=+1205.594306216" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.851596 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2d80b54-e0f6-415a-9bab-1cff8e208efa-operator-scripts\") pod \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\" (UID: \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\") " Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.851897 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dsk9\" (UniqueName: \"kubernetes.io/projected/f2d80b54-e0f6-415a-9bab-1cff8e208efa-kube-api-access-4dsk9\") pod \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\" (UID: \"f2d80b54-e0f6-415a-9bab-1cff8e208efa\") " Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.852042 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/493558c5-9375-43a2-92e1-9f20470fc681-operator-scripts\") pod \"493558c5-9375-43a2-92e1-9f20470fc681\" (UID: \"493558c5-9375-43a2-92e1-9f20470fc681\") " Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.852162 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2h9v\" (UniqueName: \"kubernetes.io/projected/493558c5-9375-43a2-92e1-9f20470fc681-kube-api-access-z2h9v\") pod \"493558c5-9375-43a2-92e1-9f20470fc681\" (UID: \"493558c5-9375-43a2-92e1-9f20470fc681\") " Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.852243 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85484764-8eab-4bbd-8d2f-1726f287ada3-operator-scripts\") pod \"85484764-8eab-4bbd-8d2f-1726f287ada3\" (UID: \"85484764-8eab-4bbd-8d2f-1726f287ada3\") " Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.852362 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8kzf\" (UniqueName: \"kubernetes.io/projected/85484764-8eab-4bbd-8d2f-1726f287ada3-kube-api-access-d8kzf\") pod \"85484764-8eab-4bbd-8d2f-1726f287ada3\" (UID: \"85484764-8eab-4bbd-8d2f-1726f287ada3\") " Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.853981 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2d80b54-e0f6-415a-9bab-1cff8e208efa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f2d80b54-e0f6-415a-9bab-1cff8e208efa" (UID: "f2d80b54-e0f6-415a-9bab-1cff8e208efa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.855173 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85484764-8eab-4bbd-8d2f-1726f287ada3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85484764-8eab-4bbd-8d2f-1726f287ada3" (UID: "85484764-8eab-4bbd-8d2f-1726f287ada3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.855610 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/493558c5-9375-43a2-92e1-9f20470fc681-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "493558c5-9375-43a2-92e1-9f20470fc681" (UID: "493558c5-9375-43a2-92e1-9f20470fc681"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.859181 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2d80b54-e0f6-415a-9bab-1cff8e208efa-kube-api-access-4dsk9" (OuterVolumeSpecName: "kube-api-access-4dsk9") pod "f2d80b54-e0f6-415a-9bab-1cff8e208efa" (UID: "f2d80b54-e0f6-415a-9bab-1cff8e208efa"). InnerVolumeSpecName "kube-api-access-4dsk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.862262 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493558c5-9375-43a2-92e1-9f20470fc681-kube-api-access-z2h9v" (OuterVolumeSpecName: "kube-api-access-z2h9v") pod "493558c5-9375-43a2-92e1-9f20470fc681" (UID: "493558c5-9375-43a2-92e1-9f20470fc681"). InnerVolumeSpecName "kube-api-access-z2h9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.876152 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85484764-8eab-4bbd-8d2f-1726f287ada3-kube-api-access-d8kzf" (OuterVolumeSpecName: "kube-api-access-d8kzf") pod "85484764-8eab-4bbd-8d2f-1726f287ada3" (UID: "85484764-8eab-4bbd-8d2f-1726f287ada3"). InnerVolumeSpecName "kube-api-access-d8kzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.955993 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/493558c5-9375-43a2-92e1-9f20470fc681-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.956021 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2h9v\" (UniqueName: \"kubernetes.io/projected/493558c5-9375-43a2-92e1-9f20470fc681-kube-api-access-z2h9v\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.956032 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85484764-8eab-4bbd-8d2f-1726f287ada3-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.956040 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8kzf\" (UniqueName: \"kubernetes.io/projected/85484764-8eab-4bbd-8d2f-1726f287ada3-kube-api-access-d8kzf\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.956049 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2d80b54-e0f6-415a-9bab-1cff8e208efa-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:37 crc kubenswrapper[4735]: I0215 20:36:37.956057 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dsk9\" (UniqueName: \"kubernetes.io/projected/f2d80b54-e0f6-415a-9bab-1cff8e208efa-kube-api-access-4dsk9\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.696366 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3a32ed00-6692-40df-af48-ecd34af2ffd2","Type":"ContainerStarted","Data":"592eeb7b3654ee939536331d4cf988fe47d5f85a048d1615a28f28f32009879e"} Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.697923 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91998299-81cc-409a-a32d-2d2fa671d379","Type":"ContainerStarted","Data":"c4df29ed0a4f445963e3a52ac2115ba1c8087a4c466235e80c1182ddc8a0bdd7"} Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.700813 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d824-account-create-update-gmj74" Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.700809 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d824-account-create-update-gmj74" event={"ID":"f2d80b54-e0f6-415a-9bab-1cff8e208efa","Type":"ContainerDied","Data":"7222645eea1f465a186e03ef416b71868ff9eed53957eac51dcc4ba3c0f99ea7"} Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.700939 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7222645eea1f465a186e03ef416b71868ff9eed53957eac51dcc4ba3c0f99ea7" Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.718891 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerStarted","Data":"c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583"} Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.720857 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f66hh" Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.720869 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f66hh" event={"ID":"85484764-8eab-4bbd-8d2f-1726f287ada3","Type":"ContainerDied","Data":"a165350886231990f6f8a1900202ef0c5ef107626c77012da09ab5c9a593d60f"} Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.720898 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a165350886231990f6f8a1900202ef0c5ef107626c77012da09ab5c9a593d60f" Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.722843 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-33f9-account-create-update-t72d7" Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.725527 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.725510269 podStartE2EDuration="6.725510269s" podCreationTimestamp="2026-02-15 20:36:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:38.719824955 +0000 UTC m=+1206.585840588" watchObservedRunningTime="2026-02-15 20:36:38.725510269 +0000 UTC m=+1206.591525882" Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.732355 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1f3720a1-390e-457d-a4c2-d758b04a90d3","Type":"ContainerStarted","Data":"b04b56947f3a4e459f9c7b2a8d0618ccc3c732219ffb6247df1adde59841606f"} Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.732420 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 15 20:36:38 crc kubenswrapper[4735]: I0215 20:36:38.775301 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.775278887 podStartE2EDuration="5.775278887s" podCreationTimestamp="2026-02-15 20:36:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:36:38.760865527 +0000 UTC m=+1206.626881150" watchObservedRunningTime="2026-02-15 20:36:38.775278887 +0000 UTC m=+1206.641294500" Feb 15 20:36:39 crc kubenswrapper[4735]: I0215 20:36:39.663551 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:39 crc kubenswrapper[4735]: I0215 20:36:39.734001 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerStarted","Data":"7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496"} Feb 15 20:36:39 crc kubenswrapper[4735]: I0215 20:36:39.771718 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.350837499 podStartE2EDuration="7.771703052s" podCreationTimestamp="2026-02-15 20:36:32 +0000 UTC" firstStartedPulling="2026-02-15 20:36:34.945162453 +0000 UTC m=+1202.811178076" lastFinishedPulling="2026-02-15 20:36:39.366028006 +0000 UTC m=+1207.232043629" observedRunningTime="2026-02-15 20:36:39.764666151 +0000 UTC m=+1207.630681774" watchObservedRunningTime="2026-02-15 20:36:39.771703052 +0000 UTC m=+1207.637718675" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.104685 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m8x9h"] Feb 15 20:36:40 crc kubenswrapper[4735]: E0215 20:36:40.105071 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85484764-8eab-4bbd-8d2f-1726f287ada3" containerName="mariadb-database-create" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105088 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="85484764-8eab-4bbd-8d2f-1726f287ada3" containerName="mariadb-database-create" Feb 15 20:36:40 crc kubenswrapper[4735]: E0215 20:36:40.105108 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f27472c1-f67e-49e6-bd56-a36101f69c64" containerName="mariadb-account-create-update" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105115 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f27472c1-f67e-49e6-bd56-a36101f69c64" containerName="mariadb-account-create-update" Feb 15 20:36:40 crc kubenswrapper[4735]: E0215 20:36:40.105123 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493558c5-9375-43a2-92e1-9f20470fc681" containerName="mariadb-account-create-update" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105129 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="493558c5-9375-43a2-92e1-9f20470fc681" containerName="mariadb-account-create-update" Feb 15 20:36:40 crc kubenswrapper[4735]: E0215 20:36:40.105148 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8507ec7-d881-4a4a-82dc-3af56d4d53aa" containerName="mariadb-database-create" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105154 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8507ec7-d881-4a4a-82dc-3af56d4d53aa" containerName="mariadb-database-create" Feb 15 20:36:40 crc kubenswrapper[4735]: E0215 20:36:40.105162 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2d80b54-e0f6-415a-9bab-1cff8e208efa" containerName="mariadb-account-create-update" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105168 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2d80b54-e0f6-415a-9bab-1cff8e208efa" containerName="mariadb-account-create-update" Feb 15 20:36:40 crc kubenswrapper[4735]: E0215 20:36:40.105185 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48189563-f3c7-49a0-b01e-86b25b077fd5" containerName="mariadb-database-create" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105191 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="48189563-f3c7-49a0-b01e-86b25b077fd5" containerName="mariadb-database-create" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105353 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="48189563-f3c7-49a0-b01e-86b25b077fd5" containerName="mariadb-database-create" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105364 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f27472c1-f67e-49e6-bd56-a36101f69c64" containerName="mariadb-account-create-update" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105376 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8507ec7-d881-4a4a-82dc-3af56d4d53aa" containerName="mariadb-database-create" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105387 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="85484764-8eab-4bbd-8d2f-1726f287ada3" containerName="mariadb-database-create" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105399 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="493558c5-9375-43a2-92e1-9f20470fc681" containerName="mariadb-account-create-update" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.105411 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2d80b54-e0f6-415a-9bab-1cff8e208efa" containerName="mariadb-account-create-update" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.106044 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.112091 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.113178 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tldpw" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.115183 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.134257 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m8x9h"] Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.210385 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-scripts\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.210547 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xxd7\" (UniqueName: \"kubernetes.io/projected/ddc4d167-d274-40c2-9d44-78bb0446ce79-kube-api-access-2xxd7\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.210618 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-config-data\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.210787 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.312315 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-scripts\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.312415 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xxd7\" (UniqueName: \"kubernetes.io/projected/ddc4d167-d274-40c2-9d44-78bb0446ce79-kube-api-access-2xxd7\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.312443 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-config-data\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.312469 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.317705 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-config-data\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.323617 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.331656 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-scripts\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.331925 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xxd7\" (UniqueName: \"kubernetes.io/projected/ddc4d167-d274-40c2-9d44-78bb0446ce79-kube-api-access-2xxd7\") pod \"nova-cell0-conductor-db-sync-m8x9h\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.422087 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.750312 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="ceilometer-central-agent" containerID="cri-o://92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c" gracePeriod=30 Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.750886 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.751176 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="proxy-httpd" containerID="cri-o://7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496" gracePeriod=30 Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.751228 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="sg-core" containerID="cri-o://c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583" gracePeriod=30 Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.751262 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="ceilometer-notification-agent" containerID="cri-o://8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00" gracePeriod=30 Feb 15 20:36:40 crc kubenswrapper[4735]: I0215 20:36:40.943122 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m8x9h"] Feb 15 20:36:41 crc kubenswrapper[4735]: I0215 20:36:41.414363 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:41 crc kubenswrapper[4735]: I0215 20:36:41.418105 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-b7b86987-zpgc8" Feb 15 20:36:41 crc kubenswrapper[4735]: I0215 20:36:41.762284 4735 generic.go:334] "Generic (PLEG): container finished" podID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerID="7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496" exitCode=0 Feb 15 20:36:41 crc kubenswrapper[4735]: I0215 20:36:41.762533 4735 generic.go:334] "Generic (PLEG): container finished" podID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerID="c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583" exitCode=2 Feb 15 20:36:41 crc kubenswrapper[4735]: I0215 20:36:41.762543 4735 generic.go:334] "Generic (PLEG): container finished" podID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerID="8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00" exitCode=0 Feb 15 20:36:41 crc kubenswrapper[4735]: I0215 20:36:41.762360 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerDied","Data":"7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496"} Feb 15 20:36:41 crc kubenswrapper[4735]: I0215 20:36:41.762599 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerDied","Data":"c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583"} Feb 15 20:36:41 crc kubenswrapper[4735]: I0215 20:36:41.762612 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerDied","Data":"8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00"} Feb 15 20:36:41 crc kubenswrapper[4735]: I0215 20:36:41.763716 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-m8x9h" event={"ID":"ddc4d167-d274-40c2-9d44-78bb0446ce79","Type":"ContainerStarted","Data":"c36c6b47f989f6333f903e6153dc3cd70dcb6af50dde513da1001e7daaa1f227"} Feb 15 20:36:42 crc kubenswrapper[4735]: I0215 20:36:42.850633 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:42 crc kubenswrapper[4735]: I0215 20:36:42.850685 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:42 crc kubenswrapper[4735]: I0215 20:36:42.910485 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:42 crc kubenswrapper[4735]: I0215 20:36:42.929506 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:43 crc kubenswrapper[4735]: I0215 20:36:43.630221 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 15 20:36:43 crc kubenswrapper[4735]: I0215 20:36:43.630777 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 15 20:36:43 crc kubenswrapper[4735]: I0215 20:36:43.702192 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 15 20:36:43 crc kubenswrapper[4735]: I0215 20:36:43.712407 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 15 20:36:43 crc kubenswrapper[4735]: I0215 20:36:43.795111 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 15 20:36:43 crc kubenswrapper[4735]: I0215 20:36:43.795140 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:43 crc kubenswrapper[4735]: I0215 20:36:43.795151 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 15 20:36:43 crc kubenswrapper[4735]: I0215 20:36:43.795482 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.297938 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.388867 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxshk\" (UniqueName: \"kubernetes.io/projected/0bcd5cef-2249-4a58-ac25-db85ed772900-kube-api-access-vxshk\") pod \"0bcd5cef-2249-4a58-ac25-db85ed772900\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.388921 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-config-data\") pod \"0bcd5cef-2249-4a58-ac25-db85ed772900\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.388953 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-scripts\") pod \"0bcd5cef-2249-4a58-ac25-db85ed772900\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.389021 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-run-httpd\") pod \"0bcd5cef-2249-4a58-ac25-db85ed772900\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.389060 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-sg-core-conf-yaml\") pod \"0bcd5cef-2249-4a58-ac25-db85ed772900\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.389586 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0bcd5cef-2249-4a58-ac25-db85ed772900" (UID: "0bcd5cef-2249-4a58-ac25-db85ed772900"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.389695 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-combined-ca-bundle\") pod \"0bcd5cef-2249-4a58-ac25-db85ed772900\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.390018 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-log-httpd\") pod \"0bcd5cef-2249-4a58-ac25-db85ed772900\" (UID: \"0bcd5cef-2249-4a58-ac25-db85ed772900\") " Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.390403 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.390614 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0bcd5cef-2249-4a58-ac25-db85ed772900" (UID: "0bcd5cef-2249-4a58-ac25-db85ed772900"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.399161 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-scripts" (OuterVolumeSpecName: "scripts") pod "0bcd5cef-2249-4a58-ac25-db85ed772900" (UID: "0bcd5cef-2249-4a58-ac25-db85ed772900"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.431065 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bcd5cef-2249-4a58-ac25-db85ed772900-kube-api-access-vxshk" (OuterVolumeSpecName: "kube-api-access-vxshk") pod "0bcd5cef-2249-4a58-ac25-db85ed772900" (UID: "0bcd5cef-2249-4a58-ac25-db85ed772900"). InnerVolumeSpecName "kube-api-access-vxshk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.435701 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0bcd5cef-2249-4a58-ac25-db85ed772900" (UID: "0bcd5cef-2249-4a58-ac25-db85ed772900"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.492229 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.492258 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bcd5cef-2249-4a58-ac25-db85ed772900-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.492266 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxshk\" (UniqueName: \"kubernetes.io/projected/0bcd5cef-2249-4a58-ac25-db85ed772900-kube-api-access-vxshk\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.492277 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.500086 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bcd5cef-2249-4a58-ac25-db85ed772900" (UID: "0bcd5cef-2249-4a58-ac25-db85ed772900"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.596214 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.628466 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-config-data" (OuterVolumeSpecName: "config-data") pod "0bcd5cef-2249-4a58-ac25-db85ed772900" (UID: "0bcd5cef-2249-4a58-ac25-db85ed772900"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.697654 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bcd5cef-2249-4a58-ac25-db85ed772900-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.778838 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.812326 4735 generic.go:334] "Generic (PLEG): container finished" podID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerID="92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c" exitCode=0 Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.813261 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.815106 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerDied","Data":"92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c"} Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.815161 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bcd5cef-2249-4a58-ac25-db85ed772900","Type":"ContainerDied","Data":"ae59437b6f6cb0d10a42840eb070de10e541f029ef61d68a312b5c32626f992f"} Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.815185 4735 scope.go:117] "RemoveContainer" containerID="7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.847751 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c647fbf5b-7zs9r" podUID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.856657 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.869022 4735 scope.go:117] "RemoveContainer" containerID="c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.880828 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.915573 4735 scope.go:117] "RemoveContainer" containerID="8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00" Feb 15 20:36:44 crc kubenswrapper[4735]: I0215 20:36:44.966374 4735 scope.go:117] "RemoveContainer" containerID="92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.014044 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" path="/var/lib/kubelet/pods/0bcd5cef-2249-4a58-ac25-db85ed772900/volumes" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.014812 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:45 crc kubenswrapper[4735]: E0215 20:36:45.015089 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="ceilometer-notification-agent" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.015105 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="ceilometer-notification-agent" Feb 15 20:36:45 crc kubenswrapper[4735]: E0215 20:36:45.015115 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="sg-core" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.015122 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="sg-core" Feb 15 20:36:45 crc kubenswrapper[4735]: E0215 20:36:45.015158 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="ceilometer-central-agent" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.015166 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="ceilometer-central-agent" Feb 15 20:36:45 crc kubenswrapper[4735]: E0215 20:36:45.015193 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="proxy-httpd" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.015199 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="proxy-httpd" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.015363 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="proxy-httpd" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.015379 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="ceilometer-notification-agent" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.015393 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="ceilometer-central-agent" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.015407 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bcd5cef-2249-4a58-ac25-db85ed772900" containerName="sg-core" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.027556 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.027663 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.028103 4735 scope.go:117] "RemoveContainer" containerID="7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496" Feb 15 20:36:45 crc kubenswrapper[4735]: E0215 20:36:45.029271 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496\": container with ID starting with 7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496 not found: ID does not exist" containerID="7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.029337 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496"} err="failed to get container status \"7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496\": rpc error: code = NotFound desc = could not find container \"7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496\": container with ID starting with 7495b8ae78dfa43f353b740a3e2600b60775852b00ddab4b7ffe6ec89f351496 not found: ID does not exist" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.029366 4735 scope.go:117] "RemoveContainer" containerID="c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583" Feb 15 20:36:45 crc kubenswrapper[4735]: E0215 20:36:45.029725 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583\": container with ID starting with c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583 not found: ID does not exist" containerID="c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.029755 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583"} err="failed to get container status \"c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583\": rpc error: code = NotFound desc = could not find container \"c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583\": container with ID starting with c688bbdcf17e5cdaec48cb394f14a7727743cc409e53233c1a5e807be4983583 not found: ID does not exist" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.029773 4735 scope.go:117] "RemoveContainer" containerID="8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00" Feb 15 20:36:45 crc kubenswrapper[4735]: E0215 20:36:45.030096 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00\": container with ID starting with 8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00 not found: ID does not exist" containerID="8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.030122 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00"} err="failed to get container status \"8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00\": rpc error: code = NotFound desc = could not find container \"8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00\": container with ID starting with 8f8728df6fd95243a732336e0f782c0cf6a85c02a0a9e692d0ad43fa358e2c00 not found: ID does not exist" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.030140 4735 scope.go:117] "RemoveContainer" containerID="92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c" Feb 15 20:36:45 crc kubenswrapper[4735]: E0215 20:36:45.030438 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c\": container with ID starting with 92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c not found: ID does not exist" containerID="92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.030463 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c"} err="failed to get container status \"92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c\": rpc error: code = NotFound desc = could not find container \"92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c\": container with ID starting with 92c2aef1f382dca3ff0616098715dcb0ffe7606b1ac6b0f955e8f8661764b39c not found: ID does not exist" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.033985 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.034491 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.106847 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.106920 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-config-data\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.107108 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj5cr\" (UniqueName: \"kubernetes.io/projected/f0390345-ed26-40f9-ab6a-63d1a0b46d21-kube-api-access-cj5cr\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.107155 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-run-httpd\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.107263 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.107304 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-log-httpd\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.107425 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-scripts\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.208604 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj5cr\" (UniqueName: \"kubernetes.io/projected/f0390345-ed26-40f9-ab6a-63d1a0b46d21-kube-api-access-cj5cr\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.209477 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-run-httpd\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.209590 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.209688 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-log-httpd\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.209779 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-scripts\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.209856 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.209961 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-config-data\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.210818 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-log-httpd\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.212094 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-run-httpd\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.217932 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-config-data\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.222415 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.223765 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.238718 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-scripts\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.243542 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj5cr\" (UniqueName: \"kubernetes.io/projected/f0390345-ed26-40f9-ab6a-63d1a0b46d21-kube-api-access-cj5cr\") pod \"ceilometer-0\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.347420 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.822804 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.823075 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.822839 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:36:45 crc kubenswrapper[4735]: I0215 20:36:45.823251 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:36:46 crc kubenswrapper[4735]: I0215 20:36:46.696119 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:48 crc kubenswrapper[4735]: I0215 20:36:48.018705 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 15 20:36:48 crc kubenswrapper[4735]: I0215 20:36:48.329508 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:48 crc kubenswrapper[4735]: I0215 20:36:48.329617 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:36:48 crc kubenswrapper[4735]: I0215 20:36:48.330241 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 15 20:36:48 crc kubenswrapper[4735]: I0215 20:36:48.423309 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 15 20:36:48 crc kubenswrapper[4735]: I0215 20:36:48.423604 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 15 20:36:48 crc kubenswrapper[4735]: I0215 20:36:48.593569 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 15 20:36:50 crc kubenswrapper[4735]: I0215 20:36:50.683151 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="1f3720a1-390e-457d-a4c2-d758b04a90d3" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.186:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 15 20:36:54 crc kubenswrapper[4735]: I0215 20:36:54.778513 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:36:54 crc kubenswrapper[4735]: I0215 20:36:54.847407 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c647fbf5b-7zs9r" podUID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 15 20:36:55 crc kubenswrapper[4735]: I0215 20:36:55.042506 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:36:55 crc kubenswrapper[4735]: W0215 20:36:55.523754 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0390345_ed26_40f9_ab6a_63d1a0b46d21.slice/crio-d4fd08b98e5b8738039b02b93a1c5fcc668baddd7d097c44c3918466ec8a6a8d WatchSource:0}: Error finding container d4fd08b98e5b8738039b02b93a1c5fcc668baddd7d097c44c3918466ec8a6a8d: Status 404 returned error can't find the container with id d4fd08b98e5b8738039b02b93a1c5fcc668baddd7d097c44c3918466ec8a6a8d Feb 15 20:36:55 crc kubenswrapper[4735]: I0215 20:36:55.953154 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-m8x9h" event={"ID":"ddc4d167-d274-40c2-9d44-78bb0446ce79","Type":"ContainerStarted","Data":"a401b2f954dce5c04bfb89229f50e102b6c7f904370f35a0bce6362e666785ee"} Feb 15 20:36:55 crc kubenswrapper[4735]: I0215 20:36:55.955929 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerStarted","Data":"d4fd08b98e5b8738039b02b93a1c5fcc668baddd7d097c44c3918466ec8a6a8d"} Feb 15 20:36:55 crc kubenswrapper[4735]: I0215 20:36:55.976217 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-m8x9h" podStartSLOduration=1.341559856 podStartE2EDuration="15.976198859s" podCreationTimestamp="2026-02-15 20:36:40 +0000 UTC" firstStartedPulling="2026-02-15 20:36:40.952573331 +0000 UTC m=+1208.818588954" lastFinishedPulling="2026-02-15 20:36:55.587212344 +0000 UTC m=+1223.453227957" observedRunningTime="2026-02-15 20:36:55.968967273 +0000 UTC m=+1223.834982896" watchObservedRunningTime="2026-02-15 20:36:55.976198859 +0000 UTC m=+1223.842214482" Feb 15 20:36:56 crc kubenswrapper[4735]: I0215 20:36:56.966701 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerStarted","Data":"84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719"} Feb 15 20:36:57 crc kubenswrapper[4735]: I0215 20:36:57.977609 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerStarted","Data":"c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb"} Feb 15 20:36:57 crc kubenswrapper[4735]: I0215 20:36:57.977901 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerStarted","Data":"579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2"} Feb 15 20:36:59 crc kubenswrapper[4735]: I0215 20:36:59.994554 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerStarted","Data":"a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69"} Feb 15 20:36:59 crc kubenswrapper[4735]: I0215 20:36:59.995144 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 15 20:36:59 crc kubenswrapper[4735]: I0215 20:36:59.995037 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="proxy-httpd" containerID="cri-o://a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69" gracePeriod=30 Feb 15 20:36:59 crc kubenswrapper[4735]: I0215 20:36:59.994740 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="ceilometer-central-agent" containerID="cri-o://84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719" gracePeriod=30 Feb 15 20:36:59 crc kubenswrapper[4735]: I0215 20:36:59.995053 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="sg-core" containerID="cri-o://c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb" gracePeriod=30 Feb 15 20:36:59 crc kubenswrapper[4735]: I0215 20:36:59.995082 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="ceilometer-notification-agent" containerID="cri-o://579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2" gracePeriod=30 Feb 15 20:37:00 crc kubenswrapper[4735]: I0215 20:37:00.030112 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=12.525223186 podStartE2EDuration="16.030088083s" podCreationTimestamp="2026-02-15 20:36:44 +0000 UTC" firstStartedPulling="2026-02-15 20:36:55.528266258 +0000 UTC m=+1223.394281871" lastFinishedPulling="2026-02-15 20:36:59.033131145 +0000 UTC m=+1226.899146768" observedRunningTime="2026-02-15 20:37:00.022272022 +0000 UTC m=+1227.888287645" watchObservedRunningTime="2026-02-15 20:37:00.030088083 +0000 UTC m=+1227.896103706" Feb 15 20:37:01 crc kubenswrapper[4735]: I0215 20:37:01.006095 4735 generic.go:334] "Generic (PLEG): container finished" podID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerID="a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69" exitCode=0 Feb 15 20:37:01 crc kubenswrapper[4735]: I0215 20:37:01.006438 4735 generic.go:334] "Generic (PLEG): container finished" podID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerID="c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb" exitCode=2 Feb 15 20:37:01 crc kubenswrapper[4735]: I0215 20:37:01.006188 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerDied","Data":"a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69"} Feb 15 20:37:01 crc kubenswrapper[4735]: I0215 20:37:01.006488 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerDied","Data":"c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb"} Feb 15 20:37:01 crc kubenswrapper[4735]: I0215 20:37:01.006501 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerDied","Data":"579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2"} Feb 15 20:37:01 crc kubenswrapper[4735]: I0215 20:37:01.006450 4735 generic.go:334] "Generic (PLEG): container finished" podID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerID="579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2" exitCode=0 Feb 15 20:37:04 crc kubenswrapper[4735]: I0215 20:37:04.778100 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:37:04 crc kubenswrapper[4735]: I0215 20:37:04.778730 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:37:04 crc kubenswrapper[4735]: I0215 20:37:04.779537 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"86ffd054acc51bbf8b9d87f36ab9440eeb9c687baf017ae70a357fa95727f187"} pod="openstack/horizon-5d466d479d-wdxj6" containerMessage="Container horizon failed startup probe, will be restarted" Feb 15 20:37:04 crc kubenswrapper[4735]: I0215 20:37:04.779579 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" containerID="cri-o://86ffd054acc51bbf8b9d87f36ab9440eeb9c687baf017ae70a357fa95727f187" gracePeriod=30 Feb 15 20:37:04 crc kubenswrapper[4735]: I0215 20:37:04.850126 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c647fbf5b-7zs9r" podUID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 15 20:37:04 crc kubenswrapper[4735]: I0215 20:37:04.850239 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:37:04 crc kubenswrapper[4735]: I0215 20:37:04.851378 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"dfabab6e6193ecc85b9a47a5ec99565c4c3f90470cab4a3b43a98447f3711f35"} pod="openstack/horizon-c647fbf5b-7zs9r" containerMessage="Container horizon failed startup probe, will be restarted" Feb 15 20:37:04 crc kubenswrapper[4735]: I0215 20:37:04.851420 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c647fbf5b-7zs9r" podUID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerName="horizon" containerID="cri-o://dfabab6e6193ecc85b9a47a5ec99565c4c3f90470cab4a3b43a98447f3711f35" gracePeriod=30 Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.088435 4735 generic.go:334] "Generic (PLEG): container finished" podID="ddc4d167-d274-40c2-9d44-78bb0446ce79" containerID="a401b2f954dce5c04bfb89229f50e102b6c7f904370f35a0bce6362e666785ee" exitCode=0 Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.088625 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-m8x9h" event={"ID":"ddc4d167-d274-40c2-9d44-78bb0446ce79","Type":"ContainerDied","Data":"a401b2f954dce5c04bfb89229f50e102b6c7f904370f35a0bce6362e666785ee"} Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.706472 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.903386 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-combined-ca-bundle\") pod \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.903650 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-config-data\") pod \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.903744 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-scripts\") pod \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.903851 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj5cr\" (UniqueName: \"kubernetes.io/projected/f0390345-ed26-40f9-ab6a-63d1a0b46d21-kube-api-access-cj5cr\") pod \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.903973 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-log-httpd\") pod \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.904602 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-sg-core-conf-yaml\") pod \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.905062 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-run-httpd\") pod \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\" (UID: \"f0390345-ed26-40f9-ab6a-63d1a0b46d21\") " Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.904533 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f0390345-ed26-40f9-ab6a-63d1a0b46d21" (UID: "f0390345-ed26-40f9-ab6a-63d1a0b46d21"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.906224 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f0390345-ed26-40f9-ab6a-63d1a0b46d21" (UID: "f0390345-ed26-40f9-ab6a-63d1a0b46d21"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.906696 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.906798 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0390345-ed26-40f9-ab6a-63d1a0b46d21-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.909357 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0390345-ed26-40f9-ab6a-63d1a0b46d21-kube-api-access-cj5cr" (OuterVolumeSpecName: "kube-api-access-cj5cr") pod "f0390345-ed26-40f9-ab6a-63d1a0b46d21" (UID: "f0390345-ed26-40f9-ab6a-63d1a0b46d21"). InnerVolumeSpecName "kube-api-access-cj5cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.910265 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-scripts" (OuterVolumeSpecName: "scripts") pod "f0390345-ed26-40f9-ab6a-63d1a0b46d21" (UID: "f0390345-ed26-40f9-ab6a-63d1a0b46d21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.933427 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f0390345-ed26-40f9-ab6a-63d1a0b46d21" (UID: "f0390345-ed26-40f9-ab6a-63d1a0b46d21"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.995847 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-config-data" (OuterVolumeSpecName: "config-data") pod "f0390345-ed26-40f9-ab6a-63d1a0b46d21" (UID: "f0390345-ed26-40f9-ab6a-63d1a0b46d21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:09 crc kubenswrapper[4735]: I0215 20:37:09.999724 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0390345-ed26-40f9-ab6a-63d1a0b46d21" (UID: "f0390345-ed26-40f9-ab6a-63d1a0b46d21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.010340 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.010372 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.010382 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.010392 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj5cr\" (UniqueName: \"kubernetes.io/projected/f0390345-ed26-40f9-ab6a-63d1a0b46d21-kube-api-access-cj5cr\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.010436 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0390345-ed26-40f9-ab6a-63d1a0b46d21-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.101720 4735 generic.go:334] "Generic (PLEG): container finished" podID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerID="84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719" exitCode=0 Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.101808 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerDied","Data":"84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719"} Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.101869 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0390345-ed26-40f9-ab6a-63d1a0b46d21","Type":"ContainerDied","Data":"d4fd08b98e5b8738039b02b93a1c5fcc668baddd7d097c44c3918466ec8a6a8d"} Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.101888 4735 scope.go:117] "RemoveContainer" containerID="a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.103611 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.155615 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.164239 4735 scope.go:117] "RemoveContainer" containerID="c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.166146 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.189094 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:10 crc kubenswrapper[4735]: E0215 20:37:10.189454 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="sg-core" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.189470 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="sg-core" Feb 15 20:37:10 crc kubenswrapper[4735]: E0215 20:37:10.189488 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="ceilometer-central-agent" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.189495 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="ceilometer-central-agent" Feb 15 20:37:10 crc kubenswrapper[4735]: E0215 20:37:10.189510 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="ceilometer-notification-agent" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.189517 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="ceilometer-notification-agent" Feb 15 20:37:10 crc kubenswrapper[4735]: E0215 20:37:10.189531 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="proxy-httpd" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.189536 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="proxy-httpd" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.189699 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="ceilometer-notification-agent" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.189714 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="ceilometer-central-agent" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.189723 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="sg-core" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.189733 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" containerName="proxy-httpd" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.191664 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.195002 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.195343 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.201255 4735 scope.go:117] "RemoveContainer" containerID="579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.207528 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.215920 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-config-data\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.216029 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-run-httpd\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.216083 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-scripts\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.216126 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-log-httpd\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.216181 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.216260 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.216341 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddfjh\" (UniqueName: \"kubernetes.io/projected/22e37e91-e16e-4413-b975-1aaf37d1f236-kube-api-access-ddfjh\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.239692 4735 scope.go:117] "RemoveContainer" containerID="84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.263924 4735 scope.go:117] "RemoveContainer" containerID="a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69" Feb 15 20:37:10 crc kubenswrapper[4735]: E0215 20:37:10.274399 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69\": container with ID starting with a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69 not found: ID does not exist" containerID="a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.274441 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69"} err="failed to get container status \"a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69\": rpc error: code = NotFound desc = could not find container \"a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69\": container with ID starting with a185033786938f27c70074c6e87ef19cfdf0587bae99b7d7031dcf6f6b75da69 not found: ID does not exist" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.274462 4735 scope.go:117] "RemoveContainer" containerID="c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb" Feb 15 20:37:10 crc kubenswrapper[4735]: E0215 20:37:10.275003 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb\": container with ID starting with c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb not found: ID does not exist" containerID="c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.275061 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb"} err="failed to get container status \"c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb\": rpc error: code = NotFound desc = could not find container \"c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb\": container with ID starting with c9c512de94252551a3244a243ddf2abca90dba49dda599a9d964e66c43f340bb not found: ID does not exist" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.275094 4735 scope.go:117] "RemoveContainer" containerID="579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2" Feb 15 20:37:10 crc kubenswrapper[4735]: E0215 20:37:10.275382 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2\": container with ID starting with 579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2 not found: ID does not exist" containerID="579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.275414 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2"} err="failed to get container status \"579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2\": rpc error: code = NotFound desc = could not find container \"579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2\": container with ID starting with 579e9d34103620a974283d65c319ff5def8456acdc043294425d6b0eb0f0c7e2 not found: ID does not exist" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.275433 4735 scope.go:117] "RemoveContainer" containerID="84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719" Feb 15 20:37:10 crc kubenswrapper[4735]: E0215 20:37:10.276160 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719\": container with ID starting with 84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719 not found: ID does not exist" containerID="84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.276189 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719"} err="failed to get container status \"84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719\": rpc error: code = NotFound desc = could not find container \"84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719\": container with ID starting with 84574d0d04a8d0897c85ebd56351ee49fd323be7da5e55cb23ee141e16346719 not found: ID does not exist" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.317810 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.317888 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddfjh\" (UniqueName: \"kubernetes.io/projected/22e37e91-e16e-4413-b975-1aaf37d1f236-kube-api-access-ddfjh\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.317910 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-config-data\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.317954 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-run-httpd\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.317988 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-scripts\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.318014 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-log-httpd\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.318050 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.319062 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-log-httpd\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.319697 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-run-httpd\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.323181 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-config-data\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.328692 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.331550 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-scripts\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.332140 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.339343 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddfjh\" (UniqueName: \"kubernetes.io/projected/22e37e91-e16e-4413-b975-1aaf37d1f236-kube-api-access-ddfjh\") pod \"ceilometer-0\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.406406 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.423227 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xxd7\" (UniqueName: \"kubernetes.io/projected/ddc4d167-d274-40c2-9d44-78bb0446ce79-kube-api-access-2xxd7\") pod \"ddc4d167-d274-40c2-9d44-78bb0446ce79\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.423310 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-config-data\") pod \"ddc4d167-d274-40c2-9d44-78bb0446ce79\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.423343 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-scripts\") pod \"ddc4d167-d274-40c2-9d44-78bb0446ce79\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.423412 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-combined-ca-bundle\") pod \"ddc4d167-d274-40c2-9d44-78bb0446ce79\" (UID: \"ddc4d167-d274-40c2-9d44-78bb0446ce79\") " Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.431090 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddc4d167-d274-40c2-9d44-78bb0446ce79-kube-api-access-2xxd7" (OuterVolumeSpecName: "kube-api-access-2xxd7") pod "ddc4d167-d274-40c2-9d44-78bb0446ce79" (UID: "ddc4d167-d274-40c2-9d44-78bb0446ce79"). InnerVolumeSpecName "kube-api-access-2xxd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.455084 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-scripts" (OuterVolumeSpecName: "scripts") pod "ddc4d167-d274-40c2-9d44-78bb0446ce79" (UID: "ddc4d167-d274-40c2-9d44-78bb0446ce79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.460618 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddc4d167-d274-40c2-9d44-78bb0446ce79" (UID: "ddc4d167-d274-40c2-9d44-78bb0446ce79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.469298 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-config-data" (OuterVolumeSpecName: "config-data") pod "ddc4d167-d274-40c2-9d44-78bb0446ce79" (UID: "ddc4d167-d274-40c2-9d44-78bb0446ce79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.507777 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.525635 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.525968 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xxd7\" (UniqueName: \"kubernetes.io/projected/ddc4d167-d274-40c2-9d44-78bb0446ce79-kube-api-access-2xxd7\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.526012 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.526020 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc4d167-d274-40c2-9d44-78bb0446ce79-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.898248 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0390345-ed26-40f9-ab6a-63d1a0b46d21" path="/var/lib/kubelet/pods/f0390345-ed26-40f9-ab6a-63d1a0b46d21/volumes" Feb 15 20:37:10 crc kubenswrapper[4735]: I0215 20:37:10.974293 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.121818 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerStarted","Data":"49967468f2ed13eb1670e30d06d26c1e5800e7f3dedd5d1f4d8cf64e3bdbd444"} Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.132575 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-m8x9h" event={"ID":"ddc4d167-d274-40c2-9d44-78bb0446ce79","Type":"ContainerDied","Data":"c36c6b47f989f6333f903e6153dc3cd70dcb6af50dde513da1001e7daaa1f227"} Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.132625 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c36c6b47f989f6333f903e6153dc3cd70dcb6af50dde513da1001e7daaa1f227" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.132734 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-m8x9h" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.238137 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 15 20:37:11 crc kubenswrapper[4735]: E0215 20:37:11.238798 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddc4d167-d274-40c2-9d44-78bb0446ce79" containerName="nova-cell0-conductor-db-sync" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.238815 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddc4d167-d274-40c2-9d44-78bb0446ce79" containerName="nova-cell0-conductor-db-sync" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.239039 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddc4d167-d274-40c2-9d44-78bb0446ce79" containerName="nova-cell0-conductor-db-sync" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.239656 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.256144 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.256463 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.256627 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tldpw" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.340126 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38fb2ae-a738-456f-b55e-65c70672dd2a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d38fb2ae-a738-456f-b55e-65c70672dd2a\") " pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.340259 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38fb2ae-a738-456f-b55e-65c70672dd2a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d38fb2ae-a738-456f-b55e-65c70672dd2a\") " pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.340317 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d94k8\" (UniqueName: \"kubernetes.io/projected/d38fb2ae-a738-456f-b55e-65c70672dd2a-kube-api-access-d94k8\") pod \"nova-cell0-conductor-0\" (UID: \"d38fb2ae-a738-456f-b55e-65c70672dd2a\") " pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.441862 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38fb2ae-a738-456f-b55e-65c70672dd2a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d38fb2ae-a738-456f-b55e-65c70672dd2a\") " pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.441957 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d94k8\" (UniqueName: \"kubernetes.io/projected/d38fb2ae-a738-456f-b55e-65c70672dd2a-kube-api-access-d94k8\") pod \"nova-cell0-conductor-0\" (UID: \"d38fb2ae-a738-456f-b55e-65c70672dd2a\") " pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.442040 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38fb2ae-a738-456f-b55e-65c70672dd2a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d38fb2ae-a738-456f-b55e-65c70672dd2a\") " pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.452638 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38fb2ae-a738-456f-b55e-65c70672dd2a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d38fb2ae-a738-456f-b55e-65c70672dd2a\") " pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.459458 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d94k8\" (UniqueName: \"kubernetes.io/projected/d38fb2ae-a738-456f-b55e-65c70672dd2a-kube-api-access-d94k8\") pod \"nova-cell0-conductor-0\" (UID: \"d38fb2ae-a738-456f-b55e-65c70672dd2a\") " pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.463369 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38fb2ae-a738-456f-b55e-65c70672dd2a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d38fb2ae-a738-456f-b55e-65c70672dd2a\") " pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:11 crc kubenswrapper[4735]: I0215 20:37:11.568326 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:12 crc kubenswrapper[4735]: I0215 20:37:12.006199 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 15 20:37:12 crc kubenswrapper[4735]: I0215 20:37:12.142127 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d38fb2ae-a738-456f-b55e-65c70672dd2a","Type":"ContainerStarted","Data":"09a3fb917e86bad1922173c7015448f14dcbbebc07f0d0e1ad22bfa7255dd015"} Feb 15 20:37:12 crc kubenswrapper[4735]: I0215 20:37:12.145048 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerStarted","Data":"d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd"} Feb 15 20:37:13 crc kubenswrapper[4735]: I0215 20:37:13.156035 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerStarted","Data":"29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931"} Feb 15 20:37:13 crc kubenswrapper[4735]: I0215 20:37:13.156484 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerStarted","Data":"9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544"} Feb 15 20:37:13 crc kubenswrapper[4735]: I0215 20:37:13.158388 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d38fb2ae-a738-456f-b55e-65c70672dd2a","Type":"ContainerStarted","Data":"799515847bd40f1a6b6ed5492af23dcd13ca977f082a35cc5a5c89c6a6c52c26"} Feb 15 20:37:13 crc kubenswrapper[4735]: I0215 20:37:13.159635 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:13 crc kubenswrapper[4735]: I0215 20:37:13.183196 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.183181236 podStartE2EDuration="2.183181236s" podCreationTimestamp="2026-02-15 20:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:37:13.182687672 +0000 UTC m=+1241.048703295" watchObservedRunningTime="2026-02-15 20:37:13.183181236 +0000 UTC m=+1241.049196859" Feb 15 20:37:15 crc kubenswrapper[4735]: I0215 20:37:15.180323 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerStarted","Data":"9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f"} Feb 15 20:37:15 crc kubenswrapper[4735]: I0215 20:37:15.181897 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 15 20:37:15 crc kubenswrapper[4735]: I0215 20:37:15.206122 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9103530050000002 podStartE2EDuration="5.206094828s" podCreationTimestamp="2026-02-15 20:37:10 +0000 UTC" firstStartedPulling="2026-02-15 20:37:10.982448057 +0000 UTC m=+1238.848463700" lastFinishedPulling="2026-02-15 20:37:14.2781899 +0000 UTC m=+1242.144205523" observedRunningTime="2026-02-15 20:37:15.200663281 +0000 UTC m=+1243.066678904" watchObservedRunningTime="2026-02-15 20:37:15.206094828 +0000 UTC m=+1243.072110461" Feb 15 20:37:17 crc kubenswrapper[4735]: I0215 20:37:17.043646 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:18 crc kubenswrapper[4735]: I0215 20:37:18.202827 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="ceilometer-notification-agent" containerID="cri-o://9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544" gracePeriod=30 Feb 15 20:37:18 crc kubenswrapper[4735]: I0215 20:37:18.202837 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="proxy-httpd" containerID="cri-o://9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f" gracePeriod=30 Feb 15 20:37:18 crc kubenswrapper[4735]: I0215 20:37:18.202844 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="sg-core" containerID="cri-o://29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931" gracePeriod=30 Feb 15 20:37:18 crc kubenswrapper[4735]: I0215 20:37:18.203002 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="ceilometer-central-agent" containerID="cri-o://d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd" gracePeriod=30 Feb 15 20:37:19 crc kubenswrapper[4735]: I0215 20:37:19.229073 4735 generic.go:334] "Generic (PLEG): container finished" podID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerID="9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f" exitCode=0 Feb 15 20:37:19 crc kubenswrapper[4735]: I0215 20:37:19.229354 4735 generic.go:334] "Generic (PLEG): container finished" podID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerID="29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931" exitCode=2 Feb 15 20:37:19 crc kubenswrapper[4735]: I0215 20:37:19.229364 4735 generic.go:334] "Generic (PLEG): container finished" podID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerID="9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544" exitCode=0 Feb 15 20:37:19 crc kubenswrapper[4735]: I0215 20:37:19.229150 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerDied","Data":"9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f"} Feb 15 20:37:19 crc kubenswrapper[4735]: I0215 20:37:19.229399 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerDied","Data":"29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931"} Feb 15 20:37:19 crc kubenswrapper[4735]: I0215 20:37:19.229413 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerDied","Data":"9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544"} Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.141137 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.242824 4735 generic.go:334] "Generic (PLEG): container finished" podID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerID="d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd" exitCode=0 Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.242869 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerDied","Data":"d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd"} Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.242894 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"22e37e91-e16e-4413-b975-1aaf37d1f236","Type":"ContainerDied","Data":"49967468f2ed13eb1670e30d06d26c1e5800e7f3dedd5d1f4d8cf64e3bdbd444"} Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.242913 4735 scope.go:117] "RemoveContainer" containerID="9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.243073 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.267543 4735 scope.go:117] "RemoveContainer" containerID="29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.290041 4735 scope.go:117] "RemoveContainer" containerID="9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.307735 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddfjh\" (UniqueName: \"kubernetes.io/projected/22e37e91-e16e-4413-b975-1aaf37d1f236-kube-api-access-ddfjh\") pod \"22e37e91-e16e-4413-b975-1aaf37d1f236\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.307815 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-run-httpd\") pod \"22e37e91-e16e-4413-b975-1aaf37d1f236\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.308007 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-combined-ca-bundle\") pod \"22e37e91-e16e-4413-b975-1aaf37d1f236\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.308083 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-log-httpd\") pod \"22e37e91-e16e-4413-b975-1aaf37d1f236\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.308155 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-config-data\") pod \"22e37e91-e16e-4413-b975-1aaf37d1f236\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.308198 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-sg-core-conf-yaml\") pod \"22e37e91-e16e-4413-b975-1aaf37d1f236\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.308229 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-scripts\") pod \"22e37e91-e16e-4413-b975-1aaf37d1f236\" (UID: \"22e37e91-e16e-4413-b975-1aaf37d1f236\") " Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.309656 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "22e37e91-e16e-4413-b975-1aaf37d1f236" (UID: "22e37e91-e16e-4413-b975-1aaf37d1f236"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.310087 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "22e37e91-e16e-4413-b975-1aaf37d1f236" (UID: "22e37e91-e16e-4413-b975-1aaf37d1f236"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.312674 4735 scope.go:117] "RemoveContainer" containerID="d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.319302 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22e37e91-e16e-4413-b975-1aaf37d1f236-kube-api-access-ddfjh" (OuterVolumeSpecName: "kube-api-access-ddfjh") pod "22e37e91-e16e-4413-b975-1aaf37d1f236" (UID: "22e37e91-e16e-4413-b975-1aaf37d1f236"). InnerVolumeSpecName "kube-api-access-ddfjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.319411 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-scripts" (OuterVolumeSpecName: "scripts") pod "22e37e91-e16e-4413-b975-1aaf37d1f236" (UID: "22e37e91-e16e-4413-b975-1aaf37d1f236"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.358246 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "22e37e91-e16e-4413-b975-1aaf37d1f236" (UID: "22e37e91-e16e-4413-b975-1aaf37d1f236"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.401634 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22e37e91-e16e-4413-b975-1aaf37d1f236" (UID: "22e37e91-e16e-4413-b975-1aaf37d1f236"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.409402 4735 scope.go:117] "RemoveContainer" containerID="9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f" Feb 15 20:37:20 crc kubenswrapper[4735]: E0215 20:37:20.409774 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f\": container with ID starting with 9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f not found: ID does not exist" containerID="9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.409802 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f"} err="failed to get container status \"9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f\": rpc error: code = NotFound desc = could not find container \"9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f\": container with ID starting with 9ef7d18be0d08aea8283ce8b68cd3b550583fa3ef3d2a69c12d03e725403d22f not found: ID does not exist" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.409821 4735 scope.go:117] "RemoveContainer" containerID="29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.410976 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411013 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411025 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411038 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411050 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddfjh\" (UniqueName: \"kubernetes.io/projected/22e37e91-e16e-4413-b975-1aaf37d1f236-kube-api-access-ddfjh\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411064 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22e37e91-e16e-4413-b975-1aaf37d1f236-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:20 crc kubenswrapper[4735]: E0215 20:37:20.411168 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931\": container with ID starting with 29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931 not found: ID does not exist" containerID="29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411198 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931"} err="failed to get container status \"29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931\": rpc error: code = NotFound desc = could not find container \"29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931\": container with ID starting with 29f17c36c83a072b219e09e96ebed43b25ff724396a13a85f248d07aa3605931 not found: ID does not exist" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411227 4735 scope.go:117] "RemoveContainer" containerID="9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544" Feb 15 20:37:20 crc kubenswrapper[4735]: E0215 20:37:20.411593 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544\": container with ID starting with 9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544 not found: ID does not exist" containerID="9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411627 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544"} err="failed to get container status \"9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544\": rpc error: code = NotFound desc = could not find container \"9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544\": container with ID starting with 9f4030a118e7f8db4bfccc1d854010b28e249ca70840e37dd4c1be413f8e7544 not found: ID does not exist" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411653 4735 scope.go:117] "RemoveContainer" containerID="d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd" Feb 15 20:37:20 crc kubenswrapper[4735]: E0215 20:37:20.411970 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd\": container with ID starting with d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd not found: ID does not exist" containerID="d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.411990 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd"} err="failed to get container status \"d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd\": rpc error: code = NotFound desc = could not find container \"d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd\": container with ID starting with d90cd7794cb46f8f57c078951c0766c0022512bad6b46f1790a96dc581d9b1fd not found: ID does not exist" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.424276 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-config-data" (OuterVolumeSpecName: "config-data") pod "22e37e91-e16e-4413-b975-1aaf37d1f236" (UID: "22e37e91-e16e-4413-b975-1aaf37d1f236"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.512292 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e37e91-e16e-4413-b975-1aaf37d1f236-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.578068 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.589140 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.623389 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:20 crc kubenswrapper[4735]: E0215 20:37:20.623749 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="ceilometer-notification-agent" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.623767 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="ceilometer-notification-agent" Feb 15 20:37:20 crc kubenswrapper[4735]: E0215 20:37:20.623793 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="sg-core" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.623799 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="sg-core" Feb 15 20:37:20 crc kubenswrapper[4735]: E0215 20:37:20.623809 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="proxy-httpd" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.623816 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="proxy-httpd" Feb 15 20:37:20 crc kubenswrapper[4735]: E0215 20:37:20.623827 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="ceilometer-central-agent" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.623833 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="ceilometer-central-agent" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.624251 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="ceilometer-notification-agent" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.624281 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="ceilometer-central-agent" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.624297 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="sg-core" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.624309 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" containerName="proxy-httpd" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.625911 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.628114 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.629347 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.641337 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.717469 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.717523 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-run-httpd\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.717726 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-config-data\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.717774 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.717926 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pntln\" (UniqueName: \"kubernetes.io/projected/56d05258-834f-4c08-b9ba-c0ba716426ce-kube-api-access-pntln\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.718141 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-log-httpd\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.718180 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-scripts\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.819502 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pntln\" (UniqueName: \"kubernetes.io/projected/56d05258-834f-4c08-b9ba-c0ba716426ce-kube-api-access-pntln\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.819620 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-log-httpd\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.819652 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-scripts\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.819722 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.819747 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-run-httpd\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.819823 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-config-data\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.819847 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.820265 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-log-httpd\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.820314 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-run-httpd\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.829643 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.829758 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-scripts\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.829892 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-config-data\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.835860 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.843110 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pntln\" (UniqueName: \"kubernetes.io/projected/56d05258-834f-4c08-b9ba-c0ba716426ce-kube-api-access-pntln\") pod \"ceilometer-0\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " pod="openstack/ceilometer-0" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.912653 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22e37e91-e16e-4413-b975-1aaf37d1f236" path="/var/lib/kubelet/pods/22e37e91-e16e-4413-b975-1aaf37d1f236/volumes" Feb 15 20:37:20 crc kubenswrapper[4735]: I0215 20:37:20.963276 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:37:21 crc kubenswrapper[4735]: I0215 20:37:21.450028 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:21 crc kubenswrapper[4735]: I0215 20:37:21.606011 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.216987 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-x6l2z"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.218475 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.221555 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.221690 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.238172 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x6l2z"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.264079 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerStarted","Data":"ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046"} Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.264124 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerStarted","Data":"c32b60a2e22b3878b0ee82203f08c2bef92e643a90811b4af521bf27e1c8b0d5"} Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.345292 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-config-data\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.345622 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7vjs\" (UniqueName: \"kubernetes.io/projected/58dfdb77-8b78-4db4-9d54-6e29394adbed-kube-api-access-s7vjs\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.345818 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.345971 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-scripts\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.447568 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7vjs\" (UniqueName: \"kubernetes.io/projected/58dfdb77-8b78-4db4-9d54-6e29394adbed-kube-api-access-s7vjs\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.447878 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.448009 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-scripts\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.448132 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-config-data\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.454925 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-config-data\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.455958 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.456470 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-scripts\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.462842 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.463897 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.494600 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7vjs\" (UniqueName: \"kubernetes.io/projected/58dfdb77-8b78-4db4-9d54-6e29394adbed-kube-api-access-s7vjs\") pod \"nova-cell0-cell-mapping-x6l2z\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.500624 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.521993 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.528434 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.530252 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.535305 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.537675 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.607111 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.613607 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.653479 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656227 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq659\" (UniqueName: \"kubernetes.io/projected/15846c8f-83d1-4c21-ad42-b41f35731b42-kube-api-access-tq659\") pod \"nova-scheduler-0\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656287 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skt9q\" (UniqueName: \"kubernetes.io/projected/03940d35-d078-4daf-999a-6c66be6c5150-kube-api-access-skt9q\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656306 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-config-data\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656320 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl2zt\" (UniqueName: \"kubernetes.io/projected/a0dce5f0-51c8-42d8-9868-4d5b657135a5-kube-api-access-kl2zt\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656342 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0dce5f0-51c8-42d8-9868-4d5b657135a5-logs\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656360 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656380 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-config-data\") pod \"nova-scheduler-0\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656402 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03940d35-d078-4daf-999a-6c66be6c5150-logs\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656415 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656510 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.656547 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-config-data\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.661888 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.679268 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759315 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq659\" (UniqueName: \"kubernetes.io/projected/15846c8f-83d1-4c21-ad42-b41f35731b42-kube-api-access-tq659\") pod \"nova-scheduler-0\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759399 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skt9q\" (UniqueName: \"kubernetes.io/projected/03940d35-d078-4daf-999a-6c66be6c5150-kube-api-access-skt9q\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759425 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-config-data\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759445 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl2zt\" (UniqueName: \"kubernetes.io/projected/a0dce5f0-51c8-42d8-9868-4d5b657135a5-kube-api-access-kl2zt\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759467 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0dce5f0-51c8-42d8-9868-4d5b657135a5-logs\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759497 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759525 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-config-data\") pod \"nova-scheduler-0\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759554 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03940d35-d078-4daf-999a-6c66be6c5150-logs\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759574 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759605 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.759629 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-config-data\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.760811 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0dce5f0-51c8-42d8-9868-4d5b657135a5-logs\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.761613 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03940d35-d078-4daf-999a-6c66be6c5150-logs\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.764696 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-config-data\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.776135 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-config-data\") pod \"nova-scheduler-0\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.776826 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.785374 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.785832 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.787993 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.792486 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.794962 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.822180 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-config-data\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.828189 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl2zt\" (UniqueName: \"kubernetes.io/projected/a0dce5f0-51c8-42d8-9868-4d5b657135a5-kube-api-access-kl2zt\") pod \"nova-api-0\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.847554 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skt9q\" (UniqueName: \"kubernetes.io/projected/03940d35-d078-4daf-999a-6c66be6c5150-kube-api-access-skt9q\") pod \"nova-metadata-0\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.865527 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq659\" (UniqueName: \"kubernetes.io/projected/15846c8f-83d1-4c21-ad42-b41f35731b42-kube-api-access-tq659\") pod \"nova-scheduler-0\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.868398 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-k7hsr"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.871200 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.890985 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.917411 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.928117 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.954919 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-k7hsr"] Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.969237 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.973687 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.973729 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4b6z\" (UniqueName: \"kubernetes.io/projected/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-kube-api-access-p4b6z\") pod \"nova-cell1-novncproxy-0\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.973822 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-config\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.973845 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktqbh\" (UniqueName: \"kubernetes.io/projected/768dae67-d800-401b-b3d8-37438bc718ae-kube-api-access-ktqbh\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.973869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.973892 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.973976 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.973993 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-svc\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:22 crc kubenswrapper[4735]: I0215 20:37:22.974066 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.076357 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.076533 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.076566 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4b6z\" (UniqueName: \"kubernetes.io/projected/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-kube-api-access-p4b6z\") pod \"nova-cell1-novncproxy-0\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.077608 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.077781 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.077906 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-config\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.081482 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-config\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.081834 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktqbh\" (UniqueName: \"kubernetes.io/projected/768dae67-d800-401b-b3d8-37438bc718ae-kube-api-access-ktqbh\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.081877 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.081912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.085065 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.085091 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-svc\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.086406 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.087054 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-svc\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.090920 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.095260 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.106767 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4b6z\" (UniqueName: \"kubernetes.io/projected/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-kube-api-access-p4b6z\") pod \"nova-cell1-novncproxy-0\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.109616 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktqbh\" (UniqueName: \"kubernetes.io/projected/768dae67-d800-401b-b3d8-37438bc718ae-kube-api-access-ktqbh\") pod \"dnsmasq-dns-bccf8f775-k7hsr\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.166042 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.230822 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.313223 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerStarted","Data":"b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9"} Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.515611 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x6l2z"] Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.831994 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:37:23 crc kubenswrapper[4735]: I0215 20:37:23.854937 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.070878 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.207555 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-k7hsr"] Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.304598 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.334032 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gcj9j"] Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.335336 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.338208 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.338467 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.366556 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c","Type":"ContainerStarted","Data":"72d47f3cf985d971efafb3a36bfdf112c9ec5fa1960e89736522210ca8b8cdb3"} Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.368405 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0dce5f0-51c8-42d8-9868-4d5b657135a5","Type":"ContainerStarted","Data":"721d00689e77f155f66284bfab1e782276371fe2f90755f338e35d33938b0040"} Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.377107 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x6l2z" event={"ID":"58dfdb77-8b78-4db4-9d54-6e29394adbed","Type":"ContainerStarted","Data":"86d3fafc558c986f2aeaf03b0f5c1adbad067a1996512df8a426bab8ee40a88e"} Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.377129 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x6l2z" event={"ID":"58dfdb77-8b78-4db4-9d54-6e29394adbed","Type":"ContainerStarted","Data":"5447024c8f872dd78868959917bd22f07a47c5d8ac977d7fc0d9901aad0d198f"} Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.411483 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gcj9j"] Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.424539 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"03940d35-d078-4daf-999a-6c66be6c5150","Type":"ContainerStarted","Data":"20fbdd76949fcb48c41875f41ca281357a9c126148cda12d73a74f1f5a6f0d1f"} Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.427656 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-scripts\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.427723 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.427750 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-config-data\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.427805 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgg9w\" (UniqueName: \"kubernetes.io/projected/487d79f2-4901-40f7-8fa4-8b38e39cfb40-kube-api-access-sgg9w\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.429974 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15846c8f-83d1-4c21-ad42-b41f35731b42","Type":"ContainerStarted","Data":"80e2e8f892927e1ef2b2816114aa04e4ee9b31a8f7b0dc70b5f08fcb0b69ba87"} Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.443110 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" event={"ID":"768dae67-d800-401b-b3d8-37438bc718ae","Type":"ContainerStarted","Data":"888af0908525aae5beeca7e3b620784497cb0fdbbdcaabb88076efc3277655c7"} Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.486030 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-x6l2z" podStartSLOduration=2.4860089690000002 podStartE2EDuration="2.486008969s" podCreationTimestamp="2026-02-15 20:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:37:24.425221184 +0000 UTC m=+1252.291236807" watchObservedRunningTime="2026-02-15 20:37:24.486008969 +0000 UTC m=+1252.352024592" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.531260 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-scripts\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.531335 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.531359 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-config-data\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.531387 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgg9w\" (UniqueName: \"kubernetes.io/projected/487d79f2-4901-40f7-8fa4-8b38e39cfb40-kube-api-access-sgg9w\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.537918 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-scripts\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.538524 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.541097 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-config-data\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.549671 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgg9w\" (UniqueName: \"kubernetes.io/projected/487d79f2-4901-40f7-8fa4-8b38e39cfb40-kube-api-access-sgg9w\") pod \"nova-cell1-conductor-db-sync-gcj9j\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:24 crc kubenswrapper[4735]: I0215 20:37:24.776034 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:25 crc kubenswrapper[4735]: I0215 20:37:25.362396 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gcj9j"] Feb 15 20:37:25 crc kubenswrapper[4735]: I0215 20:37:25.457704 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerStarted","Data":"f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13"} Feb 15 20:37:25 crc kubenswrapper[4735]: I0215 20:37:25.460485 4735 generic.go:334] "Generic (PLEG): container finished" podID="768dae67-d800-401b-b3d8-37438bc718ae" containerID="eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96" exitCode=0 Feb 15 20:37:25 crc kubenswrapper[4735]: I0215 20:37:25.460515 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" event={"ID":"768dae67-d800-401b-b3d8-37438bc718ae","Type":"ContainerDied","Data":"eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96"} Feb 15 20:37:26 crc kubenswrapper[4735]: W0215 20:37:26.196928 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod487d79f2_4901_40f7_8fa4_8b38e39cfb40.slice/crio-7fdd9443c3eeb0fc0e3519dd733000120170f8c9e0eedf8b68379ec769316125 WatchSource:0}: Error finding container 7fdd9443c3eeb0fc0e3519dd733000120170f8c9e0eedf8b68379ec769316125: Status 404 returned error can't find the container with id 7fdd9443c3eeb0fc0e3519dd733000120170f8c9e0eedf8b68379ec769316125 Feb 15 20:37:26 crc kubenswrapper[4735]: I0215 20:37:26.476019 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gcj9j" event={"ID":"487d79f2-4901-40f7-8fa4-8b38e39cfb40","Type":"ContainerStarted","Data":"7fdd9443c3eeb0fc0e3519dd733000120170f8c9e0eedf8b68379ec769316125"} Feb 15 20:37:27 crc kubenswrapper[4735]: I0215 20:37:27.244637 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 15 20:37:27 crc kubenswrapper[4735]: I0215 20:37:27.251629 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.500071 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" event={"ID":"768dae67-d800-401b-b3d8-37438bc718ae","Type":"ContainerStarted","Data":"821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2"} Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.500789 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.502990 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c","Type":"ContainerStarted","Data":"ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a"} Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.503143 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a" gracePeriod=30 Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.508739 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0dce5f0-51c8-42d8-9868-4d5b657135a5","Type":"ContainerStarted","Data":"8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825"} Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.508778 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0dce5f0-51c8-42d8-9868-4d5b657135a5","Type":"ContainerStarted","Data":"92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd"} Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.513859 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gcj9j" event={"ID":"487d79f2-4901-40f7-8fa4-8b38e39cfb40","Type":"ContainerStarted","Data":"f28691c3b1cafc1c9b500be9d06d790910e02838248da570947e81cb0b2917fd"} Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.567300 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" podStartSLOduration=6.567275315 podStartE2EDuration="6.567275315s" podCreationTimestamp="2026-02-15 20:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:37:28.52572081 +0000 UTC m=+1256.391736443" watchObservedRunningTime="2026-02-15 20:37:28.567275315 +0000 UTC m=+1256.433290938" Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.597350 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-gcj9j" podStartSLOduration=4.597331159 podStartE2EDuration="4.597331159s" podCreationTimestamp="2026-02-15 20:37:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:37:28.546117233 +0000 UTC m=+1256.412132866" watchObservedRunningTime="2026-02-15 20:37:28.597331159 +0000 UTC m=+1256.463346782" Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.603901 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.118714835 podStartE2EDuration="6.603890837s" podCreationTimestamp="2026-02-15 20:37:22 +0000 UTC" firstStartedPulling="2026-02-15 20:37:24.3342539 +0000 UTC m=+1252.200269523" lastFinishedPulling="2026-02-15 20:37:27.819429902 +0000 UTC m=+1255.685445525" observedRunningTime="2026-02-15 20:37:28.568645533 +0000 UTC m=+1256.434661156" watchObservedRunningTime="2026-02-15 20:37:28.603890837 +0000 UTC m=+1256.469906460" Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.603183 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerStarted","Data":"6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449"} Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.616402 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.901502983 podStartE2EDuration="6.616383606s" podCreationTimestamp="2026-02-15 20:37:22 +0000 UTC" firstStartedPulling="2026-02-15 20:37:24.116192784 +0000 UTC m=+1251.982208407" lastFinishedPulling="2026-02-15 20:37:27.831073407 +0000 UTC m=+1255.697089030" observedRunningTime="2026-02-15 20:37:28.593380023 +0000 UTC m=+1256.459395646" watchObservedRunningTime="2026-02-15 20:37:28.616383606 +0000 UTC m=+1256.482399229" Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.620844 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"03940d35-d078-4daf-999a-6c66be6c5150","Type":"ContainerStarted","Data":"1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81"} Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.621007 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="03940d35-d078-4daf-999a-6c66be6c5150" containerName="nova-metadata-log" containerID="cri-o://1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81" gracePeriod=30 Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.622859 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="03940d35-d078-4daf-999a-6c66be6c5150" containerName="nova-metadata-metadata" containerID="cri-o://41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed" gracePeriod=30 Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.628501 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15846c8f-83d1-4c21-ad42-b41f35731b42","Type":"ContainerStarted","Data":"04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29"} Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.668909 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.3159101509999998 podStartE2EDuration="8.668886217s" podCreationTimestamp="2026-02-15 20:37:20 +0000 UTC" firstStartedPulling="2026-02-15 20:37:21.47360682 +0000 UTC m=+1249.339622463" lastFinishedPulling="2026-02-15 20:37:27.826582906 +0000 UTC m=+1255.692598529" observedRunningTime="2026-02-15 20:37:28.642330308 +0000 UTC m=+1256.508345931" watchObservedRunningTime="2026-02-15 20:37:28.668886217 +0000 UTC m=+1256.534901840" Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.671728 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.758180481 podStartE2EDuration="6.671718504s" podCreationTimestamp="2026-02-15 20:37:22 +0000 UTC" firstStartedPulling="2026-02-15 20:37:23.858187297 +0000 UTC m=+1251.724202920" lastFinishedPulling="2026-02-15 20:37:27.77172532 +0000 UTC m=+1255.637740943" observedRunningTime="2026-02-15 20:37:28.659220125 +0000 UTC m=+1256.525235748" watchObservedRunningTime="2026-02-15 20:37:28.671718504 +0000 UTC m=+1256.537734117" Feb 15 20:37:28 crc kubenswrapper[4735]: I0215 20:37:28.693405 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7730372340000002 podStartE2EDuration="6.693387661s" podCreationTimestamp="2026-02-15 20:37:22 +0000 UTC" firstStartedPulling="2026-02-15 20:37:23.899093755 +0000 UTC m=+1251.765109378" lastFinishedPulling="2026-02-15 20:37:27.819444182 +0000 UTC m=+1255.685459805" observedRunningTime="2026-02-15 20:37:28.678876108 +0000 UTC m=+1256.544891741" watchObservedRunningTime="2026-02-15 20:37:28.693387661 +0000 UTC m=+1256.559403284" Feb 15 20:37:29 crc kubenswrapper[4735]: I0215 20:37:29.638270 4735 generic.go:334] "Generic (PLEG): container finished" podID="03940d35-d078-4daf-999a-6c66be6c5150" containerID="1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81" exitCode=143 Feb 15 20:37:29 crc kubenswrapper[4735]: I0215 20:37:29.638449 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"03940d35-d078-4daf-999a-6c66be6c5150","Type":"ContainerStarted","Data":"41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed"} Feb 15 20:37:29 crc kubenswrapper[4735]: I0215 20:37:29.638601 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"03940d35-d078-4daf-999a-6c66be6c5150","Type":"ContainerDied","Data":"1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81"} Feb 15 20:37:29 crc kubenswrapper[4735]: I0215 20:37:29.639348 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 15 20:37:32 crc kubenswrapper[4735]: I0215 20:37:32.918422 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 15 20:37:32 crc kubenswrapper[4735]: I0215 20:37:32.920834 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 15 20:37:32 crc kubenswrapper[4735]: I0215 20:37:32.929186 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 15 20:37:32 crc kubenswrapper[4735]: I0215 20:37:32.929254 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 15 20:37:32 crc kubenswrapper[4735]: I0215 20:37:32.954153 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 15 20:37:32 crc kubenswrapper[4735]: I0215 20:37:32.970249 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 15 20:37:32 crc kubenswrapper[4735]: I0215 20:37:32.970301 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.167446 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.231148 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.320118 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-sjz6k"] Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.320859 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" podUID="caed4f3d-6a00-40e6-aa12-dca360010f40" containerName="dnsmasq-dns" containerID="cri-o://231fba9713abe30ab136dbf606840fd5142e31d9fa1838e0485a397e97043226" gracePeriod=10 Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.695887 4735 generic.go:334] "Generic (PLEG): container finished" podID="58dfdb77-8b78-4db4-9d54-6e29394adbed" containerID="86d3fafc558c986f2aeaf03b0f5c1adbad067a1996512df8a426bab8ee40a88e" exitCode=0 Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.696160 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x6l2z" event={"ID":"58dfdb77-8b78-4db4-9d54-6e29394adbed","Type":"ContainerDied","Data":"86d3fafc558c986f2aeaf03b0f5c1adbad067a1996512df8a426bab8ee40a88e"} Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.707541 4735 generic.go:334] "Generic (PLEG): container finished" podID="caed4f3d-6a00-40e6-aa12-dca360010f40" containerID="231fba9713abe30ab136dbf606840fd5142e31d9fa1838e0485a397e97043226" exitCode=0 Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.707674 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" event={"ID":"caed4f3d-6a00-40e6-aa12-dca360010f40","Type":"ContainerDied","Data":"231fba9713abe30ab136dbf606840fd5142e31d9fa1838e0485a397e97043226"} Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.791152 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 15 20:37:33 crc kubenswrapper[4735]: I0215 20:37:33.980471 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.060171 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.064411 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.084318 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-swift-storage-0\") pod \"caed4f3d-6a00-40e6-aa12-dca360010f40\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.084469 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-svc\") pod \"caed4f3d-6a00-40e6-aa12-dca360010f40\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.084544 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-nb\") pod \"caed4f3d-6a00-40e6-aa12-dca360010f40\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.084637 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-sb\") pod \"caed4f3d-6a00-40e6-aa12-dca360010f40\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.084688 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-config\") pod \"caed4f3d-6a00-40e6-aa12-dca360010f40\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.084789 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjxf9\" (UniqueName: \"kubernetes.io/projected/caed4f3d-6a00-40e6-aa12-dca360010f40-kube-api-access-fjxf9\") pod \"caed4f3d-6a00-40e6-aa12-dca360010f40\" (UID: \"caed4f3d-6a00-40e6-aa12-dca360010f40\") " Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.110206 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caed4f3d-6a00-40e6-aa12-dca360010f40-kube-api-access-fjxf9" (OuterVolumeSpecName: "kube-api-access-fjxf9") pod "caed4f3d-6a00-40e6-aa12-dca360010f40" (UID: "caed4f3d-6a00-40e6-aa12-dca360010f40"). InnerVolumeSpecName "kube-api-access-fjxf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.158148 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "caed4f3d-6a00-40e6-aa12-dca360010f40" (UID: "caed4f3d-6a00-40e6-aa12-dca360010f40"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.171292 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "caed4f3d-6a00-40e6-aa12-dca360010f40" (UID: "caed4f3d-6a00-40e6-aa12-dca360010f40"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.173509 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "caed4f3d-6a00-40e6-aa12-dca360010f40" (UID: "caed4f3d-6a00-40e6-aa12-dca360010f40"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.177573 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "caed4f3d-6a00-40e6-aa12-dca360010f40" (UID: "caed4f3d-6a00-40e6-aa12-dca360010f40"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.182588 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-config" (OuterVolumeSpecName: "config") pod "caed4f3d-6a00-40e6-aa12-dca360010f40" (UID: "caed4f3d-6a00-40e6-aa12-dca360010f40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.187796 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.187821 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.187834 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.187843 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.187852 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjxf9\" (UniqueName: \"kubernetes.io/projected/caed4f3d-6a00-40e6-aa12-dca360010f40-kube-api-access-fjxf9\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.187862 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/caed4f3d-6a00-40e6-aa12-dca360010f40-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.717735 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.721023 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-sjz6k" event={"ID":"caed4f3d-6a00-40e6-aa12-dca360010f40","Type":"ContainerDied","Data":"01d135c02fa967bf93f3e2241095b19ec01d69cb083acc8596d42a54b126ccbd"} Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.721143 4735 scope.go:117] "RemoveContainer" containerID="231fba9713abe30ab136dbf606840fd5142e31d9fa1838e0485a397e97043226" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.752091 4735 scope.go:117] "RemoveContainer" containerID="9a4f845c67566578f0c82ae18b26b62798e2aae0e42305d5b9a1d3b3ce600938" Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.766956 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-sjz6k"] Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.784266 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-sjz6k"] Feb 15 20:37:34 crc kubenswrapper[4735]: I0215 20:37:34.927460 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caed4f3d-6a00-40e6-aa12-dca360010f40" path="/var/lib/kubelet/pods/caed4f3d-6a00-40e6-aa12-dca360010f40/volumes" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.107137 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.204810 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-config-data\") pod \"58dfdb77-8b78-4db4-9d54-6e29394adbed\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.204983 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-scripts\") pod \"58dfdb77-8b78-4db4-9d54-6e29394adbed\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.205049 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-combined-ca-bundle\") pod \"58dfdb77-8b78-4db4-9d54-6e29394adbed\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.205086 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7vjs\" (UniqueName: \"kubernetes.io/projected/58dfdb77-8b78-4db4-9d54-6e29394adbed-kube-api-access-s7vjs\") pod \"58dfdb77-8b78-4db4-9d54-6e29394adbed\" (UID: \"58dfdb77-8b78-4db4-9d54-6e29394adbed\") " Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.221520 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-scripts" (OuterVolumeSpecName: "scripts") pod "58dfdb77-8b78-4db4-9d54-6e29394adbed" (UID: "58dfdb77-8b78-4db4-9d54-6e29394adbed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.225564 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58dfdb77-8b78-4db4-9d54-6e29394adbed-kube-api-access-s7vjs" (OuterVolumeSpecName: "kube-api-access-s7vjs") pod "58dfdb77-8b78-4db4-9d54-6e29394adbed" (UID: "58dfdb77-8b78-4db4-9d54-6e29394adbed"). InnerVolumeSpecName "kube-api-access-s7vjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.264078 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-config-data" (OuterVolumeSpecName: "config-data") pod "58dfdb77-8b78-4db4-9d54-6e29394adbed" (UID: "58dfdb77-8b78-4db4-9d54-6e29394adbed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.280024 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58dfdb77-8b78-4db4-9d54-6e29394adbed" (UID: "58dfdb77-8b78-4db4-9d54-6e29394adbed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.312000 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.312048 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.312061 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfdb77-8b78-4db4-9d54-6e29394adbed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.312081 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7vjs\" (UniqueName: \"kubernetes.io/projected/58dfdb77-8b78-4db4-9d54-6e29394adbed-kube-api-access-s7vjs\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.757545 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x6l2z" event={"ID":"58dfdb77-8b78-4db4-9d54-6e29394adbed","Type":"ContainerDied","Data":"5447024c8f872dd78868959917bd22f07a47c5d8ac977d7fc0d9901aad0d198f"} Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.757828 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5447024c8f872dd78868959917bd22f07a47c5d8ac977d7fc0d9901aad0d198f" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.757694 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x6l2z" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.787766 4735 generic.go:334] "Generic (PLEG): container finished" podID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerID="86ffd054acc51bbf8b9d87f36ab9440eeb9c687baf017ae70a357fa95727f187" exitCode=137 Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.787830 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerDied","Data":"86ffd054acc51bbf8b9d87f36ab9440eeb9c687baf017ae70a357fa95727f187"} Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.787856 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerStarted","Data":"67b53f3ca82355aac170c86227c023b886165282b3769c53e9f598b06b47689a"} Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.787872 4735 scope.go:117] "RemoveContainer" containerID="509c549dfe02bd37ca623752f3760afb9e5b4c41da36b238f99fb78d66aac256" Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.832117 4735 generic.go:334] "Generic (PLEG): container finished" podID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerID="dfabab6e6193ecc85b9a47a5ec99565c4c3f90470cab4a3b43a98447f3711f35" exitCode=137 Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.832428 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c647fbf5b-7zs9r" event={"ID":"970d5293-b621-45e1-a1e4-8cc176c9a148","Type":"ContainerDied","Data":"dfabab6e6193ecc85b9a47a5ec99565c4c3f90470cab4a3b43a98447f3711f35"} Feb 15 20:37:35 crc kubenswrapper[4735]: I0215 20:37:35.832477 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c647fbf5b-7zs9r" event={"ID":"970d5293-b621-45e1-a1e4-8cc176c9a148","Type":"ContainerStarted","Data":"0a80e7cb63dcfaccd15d4847f21f8088363178531987f35e5c6a0255aa70534b"} Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.010155 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.010742 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-api" containerID="cri-o://8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825" gracePeriod=30 Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.011022 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-log" containerID="cri-o://92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd" gracePeriod=30 Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.051205 4735 scope.go:117] "RemoveContainer" containerID="063addf10589b9afd3c588e4171de6d504c19a3cbf44f266f86a25dfe5ca99e3" Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.056911 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.854047 4735 generic.go:334] "Generic (PLEG): container finished" podID="487d79f2-4901-40f7-8fa4-8b38e39cfb40" containerID="f28691c3b1cafc1c9b500be9d06d790910e02838248da570947e81cb0b2917fd" exitCode=0 Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.854128 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gcj9j" event={"ID":"487d79f2-4901-40f7-8fa4-8b38e39cfb40","Type":"ContainerDied","Data":"f28691c3b1cafc1c9b500be9d06d790910e02838248da570947e81cb0b2917fd"} Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.858818 4735 generic.go:334] "Generic (PLEG): container finished" podID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerID="92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd" exitCode=143 Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.858876 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0dce5f0-51c8-42d8-9868-4d5b657135a5","Type":"ContainerDied","Data":"92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd"} Feb 15 20:37:36 crc kubenswrapper[4735]: I0215 20:37:36.862053 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="15846c8f-83d1-4c21-ad42-b41f35731b42" containerName="nova-scheduler-scheduler" containerID="cri-o://04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29" gracePeriod=30 Feb 15 20:37:37 crc kubenswrapper[4735]: E0215 20:37:37.930423 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 15 20:37:37 crc kubenswrapper[4735]: E0215 20:37:37.932167 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 15 20:37:37 crc kubenswrapper[4735]: E0215 20:37:37.935556 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 15 20:37:37 crc kubenswrapper[4735]: E0215 20:37:37.935606 4735 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="15846c8f-83d1-4c21-ad42-b41f35731b42" containerName="nova-scheduler-scheduler" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.224606 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.275434 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-config-data\") pod \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.275494 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-combined-ca-bundle\") pod \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.275584 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-scripts\") pod \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.275684 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgg9w\" (UniqueName: \"kubernetes.io/projected/487d79f2-4901-40f7-8fa4-8b38e39cfb40-kube-api-access-sgg9w\") pod \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\" (UID: \"487d79f2-4901-40f7-8fa4-8b38e39cfb40\") " Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.281695 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-scripts" (OuterVolumeSpecName: "scripts") pod "487d79f2-4901-40f7-8fa4-8b38e39cfb40" (UID: "487d79f2-4901-40f7-8fa4-8b38e39cfb40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.291120 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/487d79f2-4901-40f7-8fa4-8b38e39cfb40-kube-api-access-sgg9w" (OuterVolumeSpecName: "kube-api-access-sgg9w") pod "487d79f2-4901-40f7-8fa4-8b38e39cfb40" (UID: "487d79f2-4901-40f7-8fa4-8b38e39cfb40"). InnerVolumeSpecName "kube-api-access-sgg9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.307772 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "487d79f2-4901-40f7-8fa4-8b38e39cfb40" (UID: "487d79f2-4901-40f7-8fa4-8b38e39cfb40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.314391 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-config-data" (OuterVolumeSpecName: "config-data") pod "487d79f2-4901-40f7-8fa4-8b38e39cfb40" (UID: "487d79f2-4901-40f7-8fa4-8b38e39cfb40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.378274 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgg9w\" (UniqueName: \"kubernetes.io/projected/487d79f2-4901-40f7-8fa4-8b38e39cfb40-kube-api-access-sgg9w\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.378439 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.378514 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.378572 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/487d79f2-4901-40f7-8fa4-8b38e39cfb40-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.880313 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gcj9j" event={"ID":"487d79f2-4901-40f7-8fa4-8b38e39cfb40","Type":"ContainerDied","Data":"7fdd9443c3eeb0fc0e3519dd733000120170f8c9e0eedf8b68379ec769316125"} Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.880370 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fdd9443c3eeb0fc0e3519dd733000120170f8c9e0eedf8b68379ec769316125" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.880409 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gcj9j" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.955709 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 15 20:37:38 crc kubenswrapper[4735]: E0215 20:37:38.956139 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="487d79f2-4901-40f7-8fa4-8b38e39cfb40" containerName="nova-cell1-conductor-db-sync" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.956153 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="487d79f2-4901-40f7-8fa4-8b38e39cfb40" containerName="nova-cell1-conductor-db-sync" Feb 15 20:37:38 crc kubenswrapper[4735]: E0215 20:37:38.956181 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58dfdb77-8b78-4db4-9d54-6e29394adbed" containerName="nova-manage" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.956187 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="58dfdb77-8b78-4db4-9d54-6e29394adbed" containerName="nova-manage" Feb 15 20:37:38 crc kubenswrapper[4735]: E0215 20:37:38.956203 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caed4f3d-6a00-40e6-aa12-dca360010f40" containerName="init" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.956209 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="caed4f3d-6a00-40e6-aa12-dca360010f40" containerName="init" Feb 15 20:37:38 crc kubenswrapper[4735]: E0215 20:37:38.956223 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caed4f3d-6a00-40e6-aa12-dca360010f40" containerName="dnsmasq-dns" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.956229 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="caed4f3d-6a00-40e6-aa12-dca360010f40" containerName="dnsmasq-dns" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.956411 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="58dfdb77-8b78-4db4-9d54-6e29394adbed" containerName="nova-manage" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.956425 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="487d79f2-4901-40f7-8fa4-8b38e39cfb40" containerName="nova-cell1-conductor-db-sync" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.956436 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="caed4f3d-6a00-40e6-aa12-dca360010f40" containerName="dnsmasq-dns" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.957021 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.963432 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 15 20:37:38 crc kubenswrapper[4735]: I0215 20:37:38.973490 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.089401 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943d7959-9276-4940-8783-27492dfdf13e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"943d7959-9276-4940-8783-27492dfdf13e\") " pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.089468 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/943d7959-9276-4940-8783-27492dfdf13e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"943d7959-9276-4940-8783-27492dfdf13e\") " pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.089498 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lc7d\" (UniqueName: \"kubernetes.io/projected/943d7959-9276-4940-8783-27492dfdf13e-kube-api-access-2lc7d\") pod \"nova-cell1-conductor-0\" (UID: \"943d7959-9276-4940-8783-27492dfdf13e\") " pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.191925 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943d7959-9276-4940-8783-27492dfdf13e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"943d7959-9276-4940-8783-27492dfdf13e\") " pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.192044 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/943d7959-9276-4940-8783-27492dfdf13e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"943d7959-9276-4940-8783-27492dfdf13e\") " pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.192065 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lc7d\" (UniqueName: \"kubernetes.io/projected/943d7959-9276-4940-8783-27492dfdf13e-kube-api-access-2lc7d\") pod \"nova-cell1-conductor-0\" (UID: \"943d7959-9276-4940-8783-27492dfdf13e\") " pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.197016 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943d7959-9276-4940-8783-27492dfdf13e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"943d7959-9276-4940-8783-27492dfdf13e\") " pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.199075 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/943d7959-9276-4940-8783-27492dfdf13e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"943d7959-9276-4940-8783-27492dfdf13e\") " pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.209353 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lc7d\" (UniqueName: \"kubernetes.io/projected/943d7959-9276-4940-8783-27492dfdf13e-kube-api-access-2lc7d\") pod \"nova-cell1-conductor-0\" (UID: \"943d7959-9276-4940-8783-27492dfdf13e\") " pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.276077 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.762458 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 15 20:37:39 crc kubenswrapper[4735]: I0215 20:37:39.922624 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"943d7959-9276-4940-8783-27492dfdf13e","Type":"ContainerStarted","Data":"d620b3e9121862de5377dd7a2334a47d8f9a86a188bad8bb97fcd955cbabfbf0"} Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.401198 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.428702 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-config-data\") pod \"15846c8f-83d1-4c21-ad42-b41f35731b42\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.428773 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-combined-ca-bundle\") pod \"15846c8f-83d1-4c21-ad42-b41f35731b42\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.428844 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq659\" (UniqueName: \"kubernetes.io/projected/15846c8f-83d1-4c21-ad42-b41f35731b42-kube-api-access-tq659\") pod \"15846c8f-83d1-4c21-ad42-b41f35731b42\" (UID: \"15846c8f-83d1-4c21-ad42-b41f35731b42\") " Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.435218 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15846c8f-83d1-4c21-ad42-b41f35731b42-kube-api-access-tq659" (OuterVolumeSpecName: "kube-api-access-tq659") pod "15846c8f-83d1-4c21-ad42-b41f35731b42" (UID: "15846c8f-83d1-4c21-ad42-b41f35731b42"). InnerVolumeSpecName "kube-api-access-tq659". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.472761 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15846c8f-83d1-4c21-ad42-b41f35731b42" (UID: "15846c8f-83d1-4c21-ad42-b41f35731b42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.479373 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-config-data" (OuterVolumeSpecName: "config-data") pod "15846c8f-83d1-4c21-ad42-b41f35731b42" (UID: "15846c8f-83d1-4c21-ad42-b41f35731b42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.531610 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.531637 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15846c8f-83d1-4c21-ad42-b41f35731b42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.531664 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq659\" (UniqueName: \"kubernetes.io/projected/15846c8f-83d1-4c21-ad42-b41f35731b42-kube-api-access-tq659\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.835138 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.935619 4735 generic.go:334] "Generic (PLEG): container finished" podID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerID="8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825" exitCode=0 Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.935669 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0dce5f0-51c8-42d8-9868-4d5b657135a5","Type":"ContainerDied","Data":"8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825"} Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.935703 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.935725 4735 scope.go:117] "RemoveContainer" containerID="8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.935712 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0dce5f0-51c8-42d8-9868-4d5b657135a5","Type":"ContainerDied","Data":"721d00689e77f155f66284bfab1e782276371fe2f90755f338e35d33938b0040"} Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.938091 4735 generic.go:334] "Generic (PLEG): container finished" podID="15846c8f-83d1-4c21-ad42-b41f35731b42" containerID="04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29" exitCode=0 Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.938177 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15846c8f-83d1-4c21-ad42-b41f35731b42","Type":"ContainerDied","Data":"04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29"} Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.938199 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15846c8f-83d1-4c21-ad42-b41f35731b42","Type":"ContainerDied","Data":"80e2e8f892927e1ef2b2816114aa04e4ee9b31a8f7b0dc70b5f08fcb0b69ba87"} Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.938201 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.944768 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-config-data\") pod \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.944827 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0dce5f0-51c8-42d8-9868-4d5b657135a5-logs\") pod \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.945482 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0dce5f0-51c8-42d8-9868-4d5b657135a5-logs" (OuterVolumeSpecName: "logs") pod "a0dce5f0-51c8-42d8-9868-4d5b657135a5" (UID: "a0dce5f0-51c8-42d8-9868-4d5b657135a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.946783 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl2zt\" (UniqueName: \"kubernetes.io/projected/a0dce5f0-51c8-42d8-9868-4d5b657135a5-kube-api-access-kl2zt\") pod \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.946815 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-combined-ca-bundle\") pod \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\" (UID: \"a0dce5f0-51c8-42d8-9868-4d5b657135a5\") " Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.947451 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0dce5f0-51c8-42d8-9868-4d5b657135a5-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.947500 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"943d7959-9276-4940-8783-27492dfdf13e","Type":"ContainerStarted","Data":"b48fe6b9147cc5da0e4ba0c501baa4b3a4ef0689645ac55439c3062b71f5ed3c"} Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.948500 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.950746 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0dce5f0-51c8-42d8-9868-4d5b657135a5-kube-api-access-kl2zt" (OuterVolumeSpecName: "kube-api-access-kl2zt") pod "a0dce5f0-51c8-42d8-9868-4d5b657135a5" (UID: "a0dce5f0-51c8-42d8-9868-4d5b657135a5"). InnerVolumeSpecName "kube-api-access-kl2zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.984160 4735 scope.go:117] "RemoveContainer" containerID="92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd" Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.987079 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:37:40 crc kubenswrapper[4735]: I0215 20:37:40.988312 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-config-data" (OuterVolumeSpecName: "config-data") pod "a0dce5f0-51c8-42d8-9868-4d5b657135a5" (UID: "a0dce5f0-51c8-42d8-9868-4d5b657135a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.013196 4735 scope.go:117] "RemoveContainer" containerID="8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825" Feb 15 20:37:41 crc kubenswrapper[4735]: E0215 20:37:41.014260 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825\": container with ID starting with 8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825 not found: ID does not exist" containerID="8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.014290 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825"} err="failed to get container status \"8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825\": rpc error: code = NotFound desc = could not find container \"8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825\": container with ID starting with 8668d7de4dc416d826c29ec5b994572157f2ef10a91daaf3a7f9a539726a3825 not found: ID does not exist" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.014309 4735 scope.go:117] "RemoveContainer" containerID="92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.022568 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.022738 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0dce5f0-51c8-42d8-9868-4d5b657135a5" (UID: "a0dce5f0-51c8-42d8-9868-4d5b657135a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.023055 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.023043634 podStartE2EDuration="3.023043634s" podCreationTimestamp="2026-02-15 20:37:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:37:40.978294343 +0000 UTC m=+1268.844309966" watchObservedRunningTime="2026-02-15 20:37:41.023043634 +0000 UTC m=+1268.889059257" Feb 15 20:37:41 crc kubenswrapper[4735]: E0215 20:37:41.025136 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd\": container with ID starting with 92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd not found: ID does not exist" containerID="92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.025164 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd"} err="failed to get container status \"92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd\": rpc error: code = NotFound desc = could not find container \"92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd\": container with ID starting with 92c88daaa15e524430f9e7baf581a2f6358574d10c2cf97958186a1540f434fd not found: ID does not exist" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.025185 4735 scope.go:117] "RemoveContainer" containerID="04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.042193 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:37:41 crc kubenswrapper[4735]: E0215 20:37:41.042644 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-api" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.042663 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-api" Feb 15 20:37:41 crc kubenswrapper[4735]: E0215 20:37:41.042690 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15846c8f-83d1-4c21-ad42-b41f35731b42" containerName="nova-scheduler-scheduler" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.042697 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="15846c8f-83d1-4c21-ad42-b41f35731b42" containerName="nova-scheduler-scheduler" Feb 15 20:37:41 crc kubenswrapper[4735]: E0215 20:37:41.042722 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-log" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.042727 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-log" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.042906 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="15846c8f-83d1-4c21-ad42-b41f35731b42" containerName="nova-scheduler-scheduler" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.042934 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-log" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.042960 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" containerName="nova-api-api" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.043581 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.045909 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.049096 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.050473 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.050504 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl2zt\" (UniqueName: \"kubernetes.io/projected/a0dce5f0-51c8-42d8-9868-4d5b657135a5-kube-api-access-kl2zt\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.050520 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dce5f0-51c8-42d8-9868-4d5b657135a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.060713 4735 scope.go:117] "RemoveContainer" containerID="04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29" Feb 15 20:37:41 crc kubenswrapper[4735]: E0215 20:37:41.062789 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29\": container with ID starting with 04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29 not found: ID does not exist" containerID="04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.062827 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29"} err="failed to get container status \"04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29\": rpc error: code = NotFound desc = could not find container \"04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29\": container with ID starting with 04b3abde608e98ce74141abea2900212840673ae94ef23de8f00e8bf148cbf29 not found: ID does not exist" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.151688 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-config-data\") pod \"nova-scheduler-0\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.152065 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.152165 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wh8h\" (UniqueName: \"kubernetes.io/projected/36959d4f-b1ae-491c-a24c-a437aa21ff06-kube-api-access-7wh8h\") pod \"nova-scheduler-0\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.254106 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wh8h\" (UniqueName: \"kubernetes.io/projected/36959d4f-b1ae-491c-a24c-a437aa21ff06-kube-api-access-7wh8h\") pod \"nova-scheduler-0\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.254170 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-config-data\") pod \"nova-scheduler-0\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.254234 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.257591 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.261375 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-config-data\") pod \"nova-scheduler-0\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.276730 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wh8h\" (UniqueName: \"kubernetes.io/projected/36959d4f-b1ae-491c-a24c-a437aa21ff06-kube-api-access-7wh8h\") pod \"nova-scheduler-0\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.367059 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.393808 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.403122 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.410055 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.411600 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.415633 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.425506 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.559336 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.559888 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac0e0a6-9c45-45d4-9070-45cdaaa04690-logs\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.559931 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpz22\" (UniqueName: \"kubernetes.io/projected/eac0e0a6-9c45-45d4-9070-45cdaaa04690-kube-api-access-gpz22\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.559990 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-config-data\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.661442 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.661517 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac0e0a6-9c45-45d4-9070-45cdaaa04690-logs\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.661550 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpz22\" (UniqueName: \"kubernetes.io/projected/eac0e0a6-9c45-45d4-9070-45cdaaa04690-kube-api-access-gpz22\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.661612 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-config-data\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.662638 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac0e0a6-9c45-45d4-9070-45cdaaa04690-logs\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.672176 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-config-data\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.673122 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.682455 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpz22\" (UniqueName: \"kubernetes.io/projected/eac0e0a6-9c45-45d4-9070-45cdaaa04690-kube-api-access-gpz22\") pod \"nova-api-0\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.804273 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:37:41 crc kubenswrapper[4735]: W0215 20:37:41.922193 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36959d4f_b1ae_491c_a24c_a437aa21ff06.slice/crio-9c4fe031e649cbb773ac4bb6611853d6d8523928a9bd08d277f907ac0495f46f WatchSource:0}: Error finding container 9c4fe031e649cbb773ac4bb6611853d6d8523928a9bd08d277f907ac0495f46f: Status 404 returned error can't find the container with id 9c4fe031e649cbb773ac4bb6611853d6d8523928a9bd08d277f907ac0495f46f Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.923442 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:37:41 crc kubenswrapper[4735]: I0215 20:37:41.973360 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"36959d4f-b1ae-491c-a24c-a437aa21ff06","Type":"ContainerStarted","Data":"9c4fe031e649cbb773ac4bb6611853d6d8523928a9bd08d277f907ac0495f46f"} Feb 15 20:37:42 crc kubenswrapper[4735]: I0215 20:37:42.245906 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:37:42 crc kubenswrapper[4735]: W0215 20:37:42.249412 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeac0e0a6_9c45_45d4_9070_45cdaaa04690.slice/crio-969c910743e4c6d74240690d9a75082878c4ebd601546a8f5f4e7a59335679be WatchSource:0}: Error finding container 969c910743e4c6d74240690d9a75082878c4ebd601546a8f5f4e7a59335679be: Status 404 returned error can't find the container with id 969c910743e4c6d74240690d9a75082878c4ebd601546a8f5f4e7a59335679be Feb 15 20:37:42 crc kubenswrapper[4735]: I0215 20:37:42.897204 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15846c8f-83d1-4c21-ad42-b41f35731b42" path="/var/lib/kubelet/pods/15846c8f-83d1-4c21-ad42-b41f35731b42/volumes" Feb 15 20:37:42 crc kubenswrapper[4735]: I0215 20:37:42.898053 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0dce5f0-51c8-42d8-9868-4d5b657135a5" path="/var/lib/kubelet/pods/a0dce5f0-51c8-42d8-9868-4d5b657135a5/volumes" Feb 15 20:37:42 crc kubenswrapper[4735]: I0215 20:37:42.983857 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"36959d4f-b1ae-491c-a24c-a437aa21ff06","Type":"ContainerStarted","Data":"794872a580ea452d0045b46c12bad3839f4202d8b7514046171f4f99a8403864"} Feb 15 20:37:42 crc kubenswrapper[4735]: I0215 20:37:42.985588 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eac0e0a6-9c45-45d4-9070-45cdaaa04690","Type":"ContainerStarted","Data":"8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a"} Feb 15 20:37:42 crc kubenswrapper[4735]: I0215 20:37:42.985626 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eac0e0a6-9c45-45d4-9070-45cdaaa04690","Type":"ContainerStarted","Data":"bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df"} Feb 15 20:37:42 crc kubenswrapper[4735]: I0215 20:37:42.985636 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eac0e0a6-9c45-45d4-9070-45cdaaa04690","Type":"ContainerStarted","Data":"969c910743e4c6d74240690d9a75082878c4ebd601546a8f5f4e7a59335679be"} Feb 15 20:37:43 crc kubenswrapper[4735]: I0215 20:37:43.005668 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.005649935 podStartE2EDuration="3.005649935s" podCreationTimestamp="2026-02-15 20:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:37:42.997593427 +0000 UTC m=+1270.863609050" watchObservedRunningTime="2026-02-15 20:37:43.005649935 +0000 UTC m=+1270.871665558" Feb 15 20:37:43 crc kubenswrapper[4735]: I0215 20:37:43.019881 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.01986084 podStartE2EDuration="2.01986084s" podCreationTimestamp="2026-02-15 20:37:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:37:43.018708479 +0000 UTC m=+1270.884724122" watchObservedRunningTime="2026-02-15 20:37:43.01986084 +0000 UTC m=+1270.885876473" Feb 15 20:37:44 crc kubenswrapper[4735]: I0215 20:37:44.310188 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 15 20:37:44 crc kubenswrapper[4735]: I0215 20:37:44.776970 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:37:44 crc kubenswrapper[4735]: I0215 20:37:44.777061 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:37:44 crc kubenswrapper[4735]: I0215 20:37:44.846410 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:37:44 crc kubenswrapper[4735]: I0215 20:37:44.847117 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:37:46 crc kubenswrapper[4735]: I0215 20:37:46.367557 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 15 20:37:50 crc kubenswrapper[4735]: I0215 20:37:50.976711 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 15 20:37:51 crc kubenswrapper[4735]: I0215 20:37:51.367811 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 15 20:37:51 crc kubenswrapper[4735]: I0215 20:37:51.393599 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 15 20:37:51 crc kubenswrapper[4735]: I0215 20:37:51.805063 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 15 20:37:51 crc kubenswrapper[4735]: I0215 20:37:51.805105 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 15 20:37:52 crc kubenswrapper[4735]: I0215 20:37:52.096458 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 15 20:37:52 crc kubenswrapper[4735]: I0215 20:37:52.890119 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 15 20:37:52 crc kubenswrapper[4735]: I0215 20:37:52.890117 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 15 20:37:54 crc kubenswrapper[4735]: I0215 20:37:54.779449 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:37:54 crc kubenswrapper[4735]: I0215 20:37:54.848030 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c647fbf5b-7zs9r" podUID="970d5293-b621-45e1-a1e4-8cc176c9a148" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 15 20:37:55 crc kubenswrapper[4735]: I0215 20:37:55.170854 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 15 20:37:55 crc kubenswrapper[4735]: I0215 20:37:55.171080 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="a01ba16b-aa0c-4811-8223-9f4c083fdeab" containerName="kube-state-metrics" containerID="cri-o://1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec" gracePeriod=30 Feb 15 20:37:55 crc kubenswrapper[4735]: I0215 20:37:55.729982 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 15 20:37:55 crc kubenswrapper[4735]: I0215 20:37:55.887187 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhwzq\" (UniqueName: \"kubernetes.io/projected/a01ba16b-aa0c-4811-8223-9f4c083fdeab-kube-api-access-hhwzq\") pod \"a01ba16b-aa0c-4811-8223-9f4c083fdeab\" (UID: \"a01ba16b-aa0c-4811-8223-9f4c083fdeab\") " Feb 15 20:37:55 crc kubenswrapper[4735]: I0215 20:37:55.901248 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a01ba16b-aa0c-4811-8223-9f4c083fdeab-kube-api-access-hhwzq" (OuterVolumeSpecName: "kube-api-access-hhwzq") pod "a01ba16b-aa0c-4811-8223-9f4c083fdeab" (UID: "a01ba16b-aa0c-4811-8223-9f4c083fdeab"). InnerVolumeSpecName "kube-api-access-hhwzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:55 crc kubenswrapper[4735]: I0215 20:37:55.989772 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhwzq\" (UniqueName: \"kubernetes.io/projected/a01ba16b-aa0c-4811-8223-9f4c083fdeab-kube-api-access-hhwzq\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.099761 4735 generic.go:334] "Generic (PLEG): container finished" podID="a01ba16b-aa0c-4811-8223-9f4c083fdeab" containerID="1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec" exitCode=2 Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.099827 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.099813 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a01ba16b-aa0c-4811-8223-9f4c083fdeab","Type":"ContainerDied","Data":"1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec"} Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.099885 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a01ba16b-aa0c-4811-8223-9f4c083fdeab","Type":"ContainerDied","Data":"9af9aae7e9d8e0991e8aa19dedfab167f8f3677103db2c44de93bfd6690e5a60"} Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.099909 4735 scope.go:117] "RemoveContainer" containerID="1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.118392 4735 scope.go:117] "RemoveContainer" containerID="1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec" Feb 15 20:37:56 crc kubenswrapper[4735]: E0215 20:37:56.119144 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec\": container with ID starting with 1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec not found: ID does not exist" containerID="1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.119216 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec"} err="failed to get container status \"1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec\": rpc error: code = NotFound desc = could not find container \"1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec\": container with ID starting with 1086643d416b72573ea26548925fa55a662465a15d6aa50e50cdc7027cb1cbec not found: ID does not exist" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.131182 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.141565 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.154531 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 15 20:37:56 crc kubenswrapper[4735]: E0215 20:37:56.154894 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01ba16b-aa0c-4811-8223-9f4c083fdeab" containerName="kube-state-metrics" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.154913 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01ba16b-aa0c-4811-8223-9f4c083fdeab" containerName="kube-state-metrics" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.155108 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01ba16b-aa0c-4811-8223-9f4c083fdeab" containerName="kube-state-metrics" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.155640 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.157524 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.158279 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.175980 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.297438 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.297878 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.298015 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.298058 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf4w9\" (UniqueName: \"kubernetes.io/projected/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-kube-api-access-tf4w9\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.399450 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.399541 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.399572 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf4w9\" (UniqueName: \"kubernetes.io/projected/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-kube-api-access-tf4w9\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.400374 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.404517 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.404876 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.405969 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.413998 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf4w9\" (UniqueName: \"kubernetes.io/projected/a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04-kube-api-access-tf4w9\") pod \"kube-state-metrics-0\" (UID: \"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04\") " pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.472191 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.903418 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a01ba16b-aa0c-4811-8223-9f4c083fdeab" path="/var/lib/kubelet/pods/a01ba16b-aa0c-4811-8223-9f4c083fdeab/volumes" Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.951008 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 15 20:37:56 crc kubenswrapper[4735]: W0215 20:37:56.968037 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0faafd8_ba5f_4947_b5f5_e3b6e84a7f04.slice/crio-17c246321247926814df66b3f27dbe354d3123ef8e97bd2deb14ec7c716ea82a WatchSource:0}: Error finding container 17c246321247926814df66b3f27dbe354d3123ef8e97bd2deb14ec7c716ea82a: Status 404 returned error can't find the container with id 17c246321247926814df66b3f27dbe354d3123ef8e97bd2deb14ec7c716ea82a Feb 15 20:37:56 crc kubenswrapper[4735]: I0215 20:37:56.971811 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 20:37:57 crc kubenswrapper[4735]: I0215 20:37:57.113502 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04","Type":"ContainerStarted","Data":"17c246321247926814df66b3f27dbe354d3123ef8e97bd2deb14ec7c716ea82a"} Feb 15 20:37:57 crc kubenswrapper[4735]: I0215 20:37:57.331925 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:37:57 crc kubenswrapper[4735]: I0215 20:37:57.332306 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="ceilometer-central-agent" containerID="cri-o://ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046" gracePeriod=30 Feb 15 20:37:57 crc kubenswrapper[4735]: I0215 20:37:57.332349 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="proxy-httpd" containerID="cri-o://6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449" gracePeriod=30 Feb 15 20:37:57 crc kubenswrapper[4735]: I0215 20:37:57.332533 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="ceilometer-notification-agent" containerID="cri-o://b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9" gracePeriod=30 Feb 15 20:37:57 crc kubenswrapper[4735]: I0215 20:37:57.333016 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="sg-core" containerID="cri-o://f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13" gracePeriod=30 Feb 15 20:37:58 crc kubenswrapper[4735]: I0215 20:37:58.122769 4735 generic.go:334] "Generic (PLEG): container finished" podID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerID="6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449" exitCode=0 Feb 15 20:37:58 crc kubenswrapper[4735]: I0215 20:37:58.123113 4735 generic.go:334] "Generic (PLEG): container finished" podID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerID="f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13" exitCode=2 Feb 15 20:37:58 crc kubenswrapper[4735]: I0215 20:37:58.123134 4735 generic.go:334] "Generic (PLEG): container finished" podID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerID="ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046" exitCode=0 Feb 15 20:37:58 crc kubenswrapper[4735]: I0215 20:37:58.122882 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerDied","Data":"6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449"} Feb 15 20:37:58 crc kubenswrapper[4735]: I0215 20:37:58.123235 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerDied","Data":"f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13"} Feb 15 20:37:58 crc kubenswrapper[4735]: I0215 20:37:58.123259 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerDied","Data":"ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046"} Feb 15 20:37:58 crc kubenswrapper[4735]: I0215 20:37:58.124324 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04","Type":"ContainerStarted","Data":"2d89e90e5df76e67b2f54d5ebe034d23920cc801d4b6e77718b20759296cd4b1"} Feb 15 20:37:58 crc kubenswrapper[4735]: I0215 20:37:58.124625 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 15 20:37:58 crc kubenswrapper[4735]: I0215 20:37:58.142855 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.762235031 podStartE2EDuration="2.142841439s" podCreationTimestamp="2026-02-15 20:37:56 +0000 UTC" firstStartedPulling="2026-02-15 20:37:56.971212579 +0000 UTC m=+1284.837228202" lastFinishedPulling="2026-02-15 20:37:57.351818987 +0000 UTC m=+1285.217834610" observedRunningTime="2026-02-15 20:37:58.13954269 +0000 UTC m=+1286.005558313" watchObservedRunningTime="2026-02-15 20:37:58.142841439 +0000 UTC m=+1286.008857062" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.024231 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.029741 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.061717 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-combined-ca-bundle\") pod \"03940d35-d078-4daf-999a-6c66be6c5150\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.061834 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4b6z\" (UniqueName: \"kubernetes.io/projected/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-kube-api-access-p4b6z\") pod \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.061869 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-config-data\") pod \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.061981 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03940d35-d078-4daf-999a-6c66be6c5150-logs\") pod \"03940d35-d078-4daf-999a-6c66be6c5150\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.061999 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-config-data\") pod \"03940d35-d078-4daf-999a-6c66be6c5150\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.062014 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skt9q\" (UniqueName: \"kubernetes.io/projected/03940d35-d078-4daf-999a-6c66be6c5150-kube-api-access-skt9q\") pod \"03940d35-d078-4daf-999a-6c66be6c5150\" (UID: \"03940d35-d078-4daf-999a-6c66be6c5150\") " Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.062134 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-combined-ca-bundle\") pod \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\" (UID: \"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c\") " Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.062638 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03940d35-d078-4daf-999a-6c66be6c5150-logs" (OuterVolumeSpecName: "logs") pod "03940d35-d078-4daf-999a-6c66be6c5150" (UID: "03940d35-d078-4daf-999a-6c66be6c5150"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.069016 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03940d35-d078-4daf-999a-6c66be6c5150-kube-api-access-skt9q" (OuterVolumeSpecName: "kube-api-access-skt9q") pod "03940d35-d078-4daf-999a-6c66be6c5150" (UID: "03940d35-d078-4daf-999a-6c66be6c5150"). InnerVolumeSpecName "kube-api-access-skt9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.071629 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-kube-api-access-p4b6z" (OuterVolumeSpecName: "kube-api-access-p4b6z") pod "30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c" (UID: "30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c"). InnerVolumeSpecName "kube-api-access-p4b6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.108069 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03940d35-d078-4daf-999a-6c66be6c5150" (UID: "03940d35-d078-4daf-999a-6c66be6c5150"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.108222 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c" (UID: "30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.108762 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-config-data" (OuterVolumeSpecName: "config-data") pod "03940d35-d078-4daf-999a-6c66be6c5150" (UID: "03940d35-d078-4daf-999a-6c66be6c5150"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.118722 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-config-data" (OuterVolumeSpecName: "config-data") pod "30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c" (UID: "30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.146038 4735 generic.go:334] "Generic (PLEG): container finished" podID="30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c" containerID="ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a" exitCode=137 Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.146376 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c","Type":"ContainerDied","Data":"ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a"} Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.147255 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c","Type":"ContainerDied","Data":"72d47f3cf985d971efafb3a36bfdf112c9ec5fa1960e89736522210ca8b8cdb3"} Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.147344 4735 scope.go:117] "RemoveContainer" containerID="ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.146471 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.153733 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.153774 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"03940d35-d078-4daf-999a-6c66be6c5150","Type":"ContainerDied","Data":"41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed"} Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.153878 4735 generic.go:334] "Generic (PLEG): container finished" podID="03940d35-d078-4daf-999a-6c66be6c5150" containerID="41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed" exitCode=137 Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.153926 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"03940d35-d078-4daf-999a-6c66be6c5150","Type":"ContainerDied","Data":"20fbdd76949fcb48c41875f41ca281357a9c126148cda12d73a74f1f5a6f0d1f"} Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.164784 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.164997 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.165100 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4b6z\" (UniqueName: \"kubernetes.io/projected/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-kube-api-access-p4b6z\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.165179 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.165239 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03940d35-d078-4daf-999a-6c66be6c5150-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.165754 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skt9q\" (UniqueName: \"kubernetes.io/projected/03940d35-d078-4daf-999a-6c66be6c5150-kube-api-access-skt9q\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.165889 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03940d35-d078-4daf-999a-6c66be6c5150-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.175023 4735 scope.go:117] "RemoveContainer" containerID="ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a" Feb 15 20:37:59 crc kubenswrapper[4735]: E0215 20:37:59.175646 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a\": container with ID starting with ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a not found: ID does not exist" containerID="ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.175736 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a"} err="failed to get container status \"ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a\": rpc error: code = NotFound desc = could not find container \"ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a\": container with ID starting with ad4c1a94ee3762d2164c35fae184d018c71ea2afbfeb3f82cb7543e4a071900a not found: ID does not exist" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.175789 4735 scope.go:117] "RemoveContainer" containerID="41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.194176 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.209389 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.211566 4735 scope.go:117] "RemoveContainer" containerID="1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.226692 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.241156 4735 scope.go:117] "RemoveContainer" containerID="41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed" Feb 15 20:37:59 crc kubenswrapper[4735]: E0215 20:37:59.243282 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed\": container with ID starting with 41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed not found: ID does not exist" containerID="41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.243322 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed"} err="failed to get container status \"41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed\": rpc error: code = NotFound desc = could not find container \"41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed\": container with ID starting with 41045050b4cdf77a0deb593bd62d3f7d7f5a6cb5432db91207d799f65e07ffed not found: ID does not exist" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.243343 4735 scope.go:117] "RemoveContainer" containerID="1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81" Feb 15 20:37:59 crc kubenswrapper[4735]: E0215 20:37:59.243599 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81\": container with ID starting with 1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81 not found: ID does not exist" containerID="1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.243641 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81"} err="failed to get container status \"1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81\": rpc error: code = NotFound desc = could not find container \"1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81\": container with ID starting with 1412cd03e1bdb259e4e48839311bf09d212f839c829817915d1a598f47fe9e81 not found: ID does not exist" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.244487 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.253460 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 15 20:37:59 crc kubenswrapper[4735]: E0215 20:37:59.253807 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03940d35-d078-4daf-999a-6c66be6c5150" containerName="nova-metadata-metadata" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.253824 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="03940d35-d078-4daf-999a-6c66be6c5150" containerName="nova-metadata-metadata" Feb 15 20:37:59 crc kubenswrapper[4735]: E0215 20:37:59.253850 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c" containerName="nova-cell1-novncproxy-novncproxy" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.253857 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c" containerName="nova-cell1-novncproxy-novncproxy" Feb 15 20:37:59 crc kubenswrapper[4735]: E0215 20:37:59.253867 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03940d35-d078-4daf-999a-6c66be6c5150" containerName="nova-metadata-log" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.253873 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="03940d35-d078-4daf-999a-6c66be6c5150" containerName="nova-metadata-log" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.254068 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="03940d35-d078-4daf-999a-6c66be6c5150" containerName="nova-metadata-metadata" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.254089 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c" containerName="nova-cell1-novncproxy-novncproxy" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.254108 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="03940d35-d078-4daf-999a-6c66be6c5150" containerName="nova-metadata-log" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.254663 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.256388 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.256578 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.257192 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.261886 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.263506 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.264603 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.264647 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.266995 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.267051 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.267074 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wf8s\" (UniqueName: \"kubernetes.io/projected/cc100b91-f66d-43d2-8b63-d52a3c27769f-kube-api-access-4wf8s\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.267094 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.267467 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.270510 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.286902 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.372999 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.373112 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.373140 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wf8s\" (UniqueName: \"kubernetes.io/projected/cc100b91-f66d-43d2-8b63-d52a3c27769f-kube-api-access-4wf8s\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.373161 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.373190 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fccd554-a077-4c87-931f-5c1c6040ec9e-logs\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.374340 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.374575 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.374607 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqmc5\" (UniqueName: \"kubernetes.io/projected/7fccd554-a077-4c87-931f-5c1c6040ec9e-kube-api-access-hqmc5\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.374679 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-config-data\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.374747 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.377469 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.378217 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.378485 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.379549 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc100b91-f66d-43d2-8b63-d52a3c27769f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.391347 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wf8s\" (UniqueName: \"kubernetes.io/projected/cc100b91-f66d-43d2-8b63-d52a3c27769f-kube-api-access-4wf8s\") pod \"nova-cell1-novncproxy-0\" (UID: \"cc100b91-f66d-43d2-8b63-d52a3c27769f\") " pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.477511 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fccd554-a077-4c87-931f-5c1c6040ec9e-logs\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.477575 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.477607 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.477628 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqmc5\" (UniqueName: \"kubernetes.io/projected/7fccd554-a077-4c87-931f-5c1c6040ec9e-kube-api-access-hqmc5\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.477681 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-config-data\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.478518 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fccd554-a077-4c87-931f-5c1c6040ec9e-logs\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.481051 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.482234 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.483506 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-config-data\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.499205 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqmc5\" (UniqueName: \"kubernetes.io/projected/7fccd554-a077-4c87-931f-5c1c6040ec9e-kube-api-access-hqmc5\") pod \"nova-metadata-0\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " pod="openstack/nova-metadata-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.587760 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:37:59 crc kubenswrapper[4735]: I0215 20:37:59.606512 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:38:00 crc kubenswrapper[4735]: I0215 20:38:00.043240 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:38:00 crc kubenswrapper[4735]: W0215 20:38:00.056760 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fccd554_a077_4c87_931f_5c1c6040ec9e.slice/crio-5fa80fc3609f09fdf5152c78d8f9672d570f7052abc555c787f6a60ff39b025e WatchSource:0}: Error finding container 5fa80fc3609f09fdf5152c78d8f9672d570f7052abc555c787f6a60ff39b025e: Status 404 returned error can't find the container with id 5fa80fc3609f09fdf5152c78d8f9672d570f7052abc555c787f6a60ff39b025e Feb 15 20:38:00 crc kubenswrapper[4735]: I0215 20:38:00.130376 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 15 20:38:00 crc kubenswrapper[4735]: I0215 20:38:00.172364 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cc100b91-f66d-43d2-8b63-d52a3c27769f","Type":"ContainerStarted","Data":"b37c96e6b7abca186a9e04017e2637a2374e4b62a22bb44751b3ca8a9c1569b3"} Feb 15 20:38:00 crc kubenswrapper[4735]: I0215 20:38:00.176100 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fccd554-a077-4c87-931f-5c1c6040ec9e","Type":"ContainerStarted","Data":"5fa80fc3609f09fdf5152c78d8f9672d570f7052abc555c787f6a60ff39b025e"} Feb 15 20:38:00 crc kubenswrapper[4735]: I0215 20:38:00.911549 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03940d35-d078-4daf-999a-6c66be6c5150" path="/var/lib/kubelet/pods/03940d35-d078-4daf-999a-6c66be6c5150/volumes" Feb 15 20:38:00 crc kubenswrapper[4735]: I0215 20:38:00.913059 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c" path="/var/lib/kubelet/pods/30e6f3d0-8c07-44fc-a86a-5ec1ed094b5c/volumes" Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.190291 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fccd554-a077-4c87-931f-5c1c6040ec9e","Type":"ContainerStarted","Data":"e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85"} Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.190336 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fccd554-a077-4c87-931f-5c1c6040ec9e","Type":"ContainerStarted","Data":"637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71"} Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.192598 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cc100b91-f66d-43d2-8b63-d52a3c27769f","Type":"ContainerStarted","Data":"2c3d5c6671350b82dfdd0368939c75f92240d22a9cc279671a0fb8764a7b417b"} Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.218899 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.218877842 podStartE2EDuration="2.218877842s" podCreationTimestamp="2026-02-15 20:37:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:38:01.212754787 +0000 UTC m=+1289.078770440" watchObservedRunningTime="2026-02-15 20:38:01.218877842 +0000 UTC m=+1289.084893475" Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.235807 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.23578297 podStartE2EDuration="2.23578297s" podCreationTimestamp="2026-02-15 20:37:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:38:01.228859722 +0000 UTC m=+1289.094875355" watchObservedRunningTime="2026-02-15 20:38:01.23578297 +0000 UTC m=+1289.101798583" Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.808312 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.809300 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.809924 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.810001 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.813975 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 15 20:38:01 crc kubenswrapper[4735]: I0215 20:38:01.815678 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.051754 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-892xq"] Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.060135 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.132245 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-892xq"] Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.223499 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.223590 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.223615 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.223643 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.223697 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-config\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.223826 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ws2k\" (UniqueName: \"kubernetes.io/projected/100107cc-9e89-476e-a44c-ca88b04793cc-kube-api-access-7ws2k\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.325968 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.326089 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-config\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.326174 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ws2k\" (UniqueName: \"kubernetes.io/projected/100107cc-9e89-476e-a44c-ca88b04793cc-kube-api-access-7ws2k\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.326306 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.326468 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.326496 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.326730 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.327126 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-config\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.327553 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.327727 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.327885 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.348824 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ws2k\" (UniqueName: \"kubernetes.io/projected/100107cc-9e89-476e-a44c-ca88b04793cc-kube-api-access-7ws2k\") pod \"dnsmasq-dns-cd5cbd7b9-892xq\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.385527 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.739338 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.921713 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-892xq"] Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.938876 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-log-httpd\") pod \"56d05258-834f-4c08-b9ba-c0ba716426ce\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.938925 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-sg-core-conf-yaml\") pod \"56d05258-834f-4c08-b9ba-c0ba716426ce\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.938957 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-combined-ca-bundle\") pod \"56d05258-834f-4c08-b9ba-c0ba716426ce\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.939002 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pntln\" (UniqueName: \"kubernetes.io/projected/56d05258-834f-4c08-b9ba-c0ba716426ce-kube-api-access-pntln\") pod \"56d05258-834f-4c08-b9ba-c0ba716426ce\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.939067 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-run-httpd\") pod \"56d05258-834f-4c08-b9ba-c0ba716426ce\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.939144 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-scripts\") pod \"56d05258-834f-4c08-b9ba-c0ba716426ce\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.939185 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-config-data\") pod \"56d05258-834f-4c08-b9ba-c0ba716426ce\" (UID: \"56d05258-834f-4c08-b9ba-c0ba716426ce\") " Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.940614 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "56d05258-834f-4c08-b9ba-c0ba716426ce" (UID: "56d05258-834f-4c08-b9ba-c0ba716426ce"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.940810 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "56d05258-834f-4c08-b9ba-c0ba716426ce" (UID: "56d05258-834f-4c08-b9ba-c0ba716426ce"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.954341 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56d05258-834f-4c08-b9ba-c0ba716426ce-kube-api-access-pntln" (OuterVolumeSpecName: "kube-api-access-pntln") pod "56d05258-834f-4c08-b9ba-c0ba716426ce" (UID: "56d05258-834f-4c08-b9ba-c0ba716426ce"). InnerVolumeSpecName "kube-api-access-pntln". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.959297 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-scripts" (OuterVolumeSpecName: "scripts") pod "56d05258-834f-4c08-b9ba-c0ba716426ce" (UID: "56d05258-834f-4c08-b9ba-c0ba716426ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:02 crc kubenswrapper[4735]: I0215 20:38:02.978438 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "56d05258-834f-4c08-b9ba-c0ba716426ce" (UID: "56d05258-834f-4c08-b9ba-c0ba716426ce"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.048146 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.048176 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.048186 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pntln\" (UniqueName: \"kubernetes.io/projected/56d05258-834f-4c08-b9ba-c0ba716426ce-kube-api-access-pntln\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.048195 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56d05258-834f-4c08-b9ba-c0ba716426ce-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.048203 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.052195 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56d05258-834f-4c08-b9ba-c0ba716426ce" (UID: "56d05258-834f-4c08-b9ba-c0ba716426ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.068611 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-config-data" (OuterVolumeSpecName: "config-data") pod "56d05258-834f-4c08-b9ba-c0ba716426ce" (UID: "56d05258-834f-4c08-b9ba-c0ba716426ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.150370 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.150672 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d05258-834f-4c08-b9ba-c0ba716426ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.212380 4735 generic.go:334] "Generic (PLEG): container finished" podID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerID="b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9" exitCode=0 Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.212513 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.213032 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerDied","Data":"b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9"} Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.213094 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56d05258-834f-4c08-b9ba-c0ba716426ce","Type":"ContainerDied","Data":"c32b60a2e22b3878b0ee82203f08c2bef92e643a90811b4af521bf27e1c8b0d5"} Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.213125 4735 scope.go:117] "RemoveContainer" containerID="6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.214870 4735 generic.go:334] "Generic (PLEG): container finished" podID="100107cc-9e89-476e-a44c-ca88b04793cc" containerID="2c24fedc4a4f755da0d3ea38c9cbfae50208025fdb6719a9009bcd932a613aa3" exitCode=0 Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.214902 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" event={"ID":"100107cc-9e89-476e-a44c-ca88b04793cc","Type":"ContainerDied","Data":"2c24fedc4a4f755da0d3ea38c9cbfae50208025fdb6719a9009bcd932a613aa3"} Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.214962 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" event={"ID":"100107cc-9e89-476e-a44c-ca88b04793cc","Type":"ContainerStarted","Data":"79e68b0b3d0355cea189ef133dc7745eb1004520bd8573b28cecd509118aba6b"} Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.241211 4735 scope.go:117] "RemoveContainer" containerID="f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.263668 4735 scope.go:117] "RemoveContainer" containerID="b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.277601 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.284853 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.303408 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:03 crc kubenswrapper[4735]: E0215 20:38:03.303740 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="proxy-httpd" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.303756 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="proxy-httpd" Feb 15 20:38:03 crc kubenswrapper[4735]: E0215 20:38:03.303775 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="ceilometer-central-agent" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.303782 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="ceilometer-central-agent" Feb 15 20:38:03 crc kubenswrapper[4735]: E0215 20:38:03.303794 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="ceilometer-notification-agent" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.303801 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="ceilometer-notification-agent" Feb 15 20:38:03 crc kubenswrapper[4735]: E0215 20:38:03.303825 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="sg-core" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.303831 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="sg-core" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.303998 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="ceilometer-notification-agent" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.304026 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="ceilometer-central-agent" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.304039 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="sg-core" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.304049 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" containerName="proxy-httpd" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.306114 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.309273 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.309482 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.309657 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.319174 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.359039 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.359075 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.359111 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-config-data\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.359135 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-run-httpd\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.359176 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-log-httpd\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.359213 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-scripts\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.359229 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgbdl\" (UniqueName: \"kubernetes.io/projected/f455dcec-5963-472c-9210-f111dac9c962-kube-api-access-sgbdl\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.359287 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.376483 4735 scope.go:117] "RemoveContainer" containerID="ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.460468 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-log-httpd\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.460525 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-scripts\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.460548 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgbdl\" (UniqueName: \"kubernetes.io/projected/f455dcec-5963-472c-9210-f111dac9c962-kube-api-access-sgbdl\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.460593 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.460659 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.460679 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.460705 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-config-data\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.460733 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-run-httpd\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.463423 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-log-httpd\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.463464 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-run-httpd\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.467376 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-config-data\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.468548 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.469169 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.470269 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-scripts\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.484206 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.485792 4735 scope.go:117] "RemoveContainer" containerID="6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449" Feb 15 20:38:03 crc kubenswrapper[4735]: E0215 20:38:03.487304 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449\": container with ID starting with 6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449 not found: ID does not exist" containerID="6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.487343 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449"} err="failed to get container status \"6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449\": rpc error: code = NotFound desc = could not find container \"6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449\": container with ID starting with 6babcd7d072202fbcbf3f651e045a9b61be60bee53aa919d6f4e08bf40027449 not found: ID does not exist" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.487364 4735 scope.go:117] "RemoveContainer" containerID="f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13" Feb 15 20:38:03 crc kubenswrapper[4735]: E0215 20:38:03.488938 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13\": container with ID starting with f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13 not found: ID does not exist" containerID="f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.488995 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13"} err="failed to get container status \"f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13\": rpc error: code = NotFound desc = could not find container \"f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13\": container with ID starting with f30c87e2b46eac93eb8f41ec263cd3c9d39b1e56a736b0e3f62b56746c794e13 not found: ID does not exist" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.489010 4735 scope.go:117] "RemoveContainer" containerID="b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9" Feb 15 20:38:03 crc kubenswrapper[4735]: E0215 20:38:03.492961 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9\": container with ID starting with b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9 not found: ID does not exist" containerID="b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.492992 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9"} err="failed to get container status \"b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9\": rpc error: code = NotFound desc = could not find container \"b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9\": container with ID starting with b3b086734f5e1b916de67c3bdd68826f276d1b5ae780ca3c36be4136954815a9 not found: ID does not exist" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.493010 4735 scope.go:117] "RemoveContainer" containerID="ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.493675 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgbdl\" (UniqueName: \"kubernetes.io/projected/f455dcec-5963-472c-9210-f111dac9c962-kube-api-access-sgbdl\") pod \"ceilometer-0\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " pod="openstack/ceilometer-0" Feb 15 20:38:03 crc kubenswrapper[4735]: E0215 20:38:03.495470 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046\": container with ID starting with ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046 not found: ID does not exist" containerID="ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.495518 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046"} err="failed to get container status \"ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046\": rpc error: code = NotFound desc = could not find container \"ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046\": container with ID starting with ba10d51ef97ea1dca7e1520aca1b0cba1183ab5f39d95fdce0fc898593796046 not found: ID does not exist" Feb 15 20:38:03 crc kubenswrapper[4735]: I0215 20:38:03.586897 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.107434 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.223937 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerStarted","Data":"b7f1e5e84c2be750f1d5bc6be76b34812120db7898bc186aaf08bd8a489c226c"} Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.228842 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" event={"ID":"100107cc-9e89-476e-a44c-ca88b04793cc","Type":"ContainerStarted","Data":"559c831cd9849fbdc2e511ca9fbec70fdddb93b5598fef560746a9ec1621c018"} Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.229001 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.254869 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" podStartSLOduration=2.25485146 podStartE2EDuration="2.25485146s" podCreationTimestamp="2026-02-15 20:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:38:04.243775961 +0000 UTC m=+1292.109791584" watchObservedRunningTime="2026-02-15 20:38:04.25485146 +0000 UTC m=+1292.120867083" Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.588220 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.588468 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-log" containerID="cri-o://bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df" gracePeriod=30 Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.588532 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.588547 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-api" containerID="cri-o://8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a" gracePeriod=30 Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.606750 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.606792 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 15 20:38:04 crc kubenswrapper[4735]: I0215 20:38:04.896642 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56d05258-834f-4c08-b9ba-c0ba716426ce" path="/var/lib/kubelet/pods/56d05258-834f-4c08-b9ba-c0ba716426ce/volumes" Feb 15 20:38:05 crc kubenswrapper[4735]: I0215 20:38:05.237187 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerStarted","Data":"bb17c94ca3be19f8abbd9ca1803411e503157fd7a17c2c3d5c5f9d2b32feae1e"} Feb 15 20:38:05 crc kubenswrapper[4735]: I0215 20:38:05.238883 4735 generic.go:334] "Generic (PLEG): container finished" podID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerID="bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df" exitCode=143 Feb 15 20:38:05 crc kubenswrapper[4735]: I0215 20:38:05.238985 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eac0e0a6-9c45-45d4-9070-45cdaaa04690","Type":"ContainerDied","Data":"bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df"} Feb 15 20:38:06 crc kubenswrapper[4735]: I0215 20:38:06.184583 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:06 crc kubenswrapper[4735]: I0215 20:38:06.248161 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerStarted","Data":"5d9ab85049e467b079c72760f19d754e9b1f1d542ef67118b0b6deba94769860"} Feb 15 20:38:06 crc kubenswrapper[4735]: I0215 20:38:06.489495 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 15 20:38:07 crc kubenswrapper[4735]: I0215 20:38:07.257000 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerStarted","Data":"515f4b1cee2de2e7761314202a05e648f8ddae955a63237c7a7766c612abe2dd"} Feb 15 20:38:07 crc kubenswrapper[4735]: I0215 20:38:07.807805 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:38:07 crc kubenswrapper[4735]: I0215 20:38:07.822620 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.248130 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.267434 4735 generic.go:334] "Generic (PLEG): container finished" podID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerID="8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a" exitCode=0 Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.267484 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.267511 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eac0e0a6-9c45-45d4-9070-45cdaaa04690","Type":"ContainerDied","Data":"8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a"} Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.267535 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eac0e0a6-9c45-45d4-9070-45cdaaa04690","Type":"ContainerDied","Data":"969c910743e4c6d74240690d9a75082878c4ebd601546a8f5f4e7a59335679be"} Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.267568 4735 scope.go:117] "RemoveContainer" containerID="8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.273327 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerStarted","Data":"73405b46773b7734fa6b22df5d5b79a1a40dc2c9778063dbd1480450b0d29a92"} Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.273636 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="ceilometer-central-agent" containerID="cri-o://bb17c94ca3be19f8abbd9ca1803411e503157fd7a17c2c3d5c5f9d2b32feae1e" gracePeriod=30 Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.273740 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="sg-core" containerID="cri-o://515f4b1cee2de2e7761314202a05e648f8ddae955a63237c7a7766c612abe2dd" gracePeriod=30 Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.273754 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="proxy-httpd" containerID="cri-o://73405b46773b7734fa6b22df5d5b79a1a40dc2c9778063dbd1480450b0d29a92" gracePeriod=30 Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.273763 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="ceilometer-notification-agent" containerID="cri-o://5d9ab85049e467b079c72760f19d754e9b1f1d542ef67118b0b6deba94769860" gracePeriod=30 Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.345181 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.148950954 podStartE2EDuration="5.345150321s" podCreationTimestamp="2026-02-15 20:38:03 +0000 UTC" firstStartedPulling="2026-02-15 20:38:04.107809758 +0000 UTC m=+1291.973825381" lastFinishedPulling="2026-02-15 20:38:07.304009125 +0000 UTC m=+1295.170024748" observedRunningTime="2026-02-15 20:38:08.332864798 +0000 UTC m=+1296.198880421" watchObservedRunningTime="2026-02-15 20:38:08.345150321 +0000 UTC m=+1296.211165964" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.358447 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-combined-ca-bundle\") pod \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.358554 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpz22\" (UniqueName: \"kubernetes.io/projected/eac0e0a6-9c45-45d4-9070-45cdaaa04690-kube-api-access-gpz22\") pod \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.358642 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-config-data\") pod \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.359392 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac0e0a6-9c45-45d4-9070-45cdaaa04690-logs\") pod \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\" (UID: \"eac0e0a6-9c45-45d4-9070-45cdaaa04690\") " Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.360456 4735 scope.go:117] "RemoveContainer" containerID="bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.371766 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eac0e0a6-9c45-45d4-9070-45cdaaa04690-logs" (OuterVolumeSpecName: "logs") pod "eac0e0a6-9c45-45d4-9070-45cdaaa04690" (UID: "eac0e0a6-9c45-45d4-9070-45cdaaa04690"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.391442 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eac0e0a6-9c45-45d4-9070-45cdaaa04690-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.398571 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac0e0a6-9c45-45d4-9070-45cdaaa04690-kube-api-access-gpz22" (OuterVolumeSpecName: "kube-api-access-gpz22") pod "eac0e0a6-9c45-45d4-9070-45cdaaa04690" (UID: "eac0e0a6-9c45-45d4-9070-45cdaaa04690"). InnerVolumeSpecName "kube-api-access-gpz22". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.415912 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eac0e0a6-9c45-45d4-9070-45cdaaa04690" (UID: "eac0e0a6-9c45-45d4-9070-45cdaaa04690"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.427029 4735 scope.go:117] "RemoveContainer" containerID="8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a" Feb 15 20:38:08 crc kubenswrapper[4735]: E0215 20:38:08.429444 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a\": container with ID starting with 8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a not found: ID does not exist" containerID="8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.429573 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a"} err="failed to get container status \"8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a\": rpc error: code = NotFound desc = could not find container \"8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a\": container with ID starting with 8df6d5409be44fd7b2d1d9dfa523a11d81b4a00ad588be96dd24028f073a0f9a not found: ID does not exist" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.429655 4735 scope.go:117] "RemoveContainer" containerID="bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df" Feb 15 20:38:08 crc kubenswrapper[4735]: E0215 20:38:08.434333 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df\": container with ID starting with bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df not found: ID does not exist" containerID="bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.434485 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df"} err="failed to get container status \"bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df\": rpc error: code = NotFound desc = could not find container \"bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df\": container with ID starting with bffa8bf4fe57f71ec440a410950fb566905aa14bb8bd3f47a53a1f9b61b740df not found: ID does not exist" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.461704 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-config-data" (OuterVolumeSpecName: "config-data") pod "eac0e0a6-9c45-45d4-9070-45cdaaa04690" (UID: "eac0e0a6-9c45-45d4-9070-45cdaaa04690"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.493126 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.493155 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpz22\" (UniqueName: \"kubernetes.io/projected/eac0e0a6-9c45-45d4-9070-45cdaaa04690-kube-api-access-gpz22\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.493164 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac0e0a6-9c45-45d4-9070-45cdaaa04690-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.655605 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.684478 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.706993 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:08 crc kubenswrapper[4735]: E0215 20:38:08.707368 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-log" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.707388 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-log" Feb 15 20:38:08 crc kubenswrapper[4735]: E0215 20:38:08.707400 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-api" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.707406 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-api" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.707582 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-api" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.707592 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" containerName="nova-api-log" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.709586 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.717155 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.718880 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.719108 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.719342 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.801715 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-public-tls-certs\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.802010 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-config-data\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.802183 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7n4j\" (UniqueName: \"kubernetes.io/projected/48d43446-4a96-4385-8045-58ec84e7335f-kube-api-access-f7n4j\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.802320 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.802438 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48d43446-4a96-4385-8045-58ec84e7335f-logs\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.802587 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.895756 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eac0e0a6-9c45-45d4-9070-45cdaaa04690" path="/var/lib/kubelet/pods/eac0e0a6-9c45-45d4-9070-45cdaaa04690/volumes" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.904616 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-public-tls-certs\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.904831 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-config-data\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.904912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7n4j\" (UniqueName: \"kubernetes.io/projected/48d43446-4a96-4385-8045-58ec84e7335f-kube-api-access-f7n4j\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.905017 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.905111 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48d43446-4a96-4385-8045-58ec84e7335f-logs\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.905232 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.906637 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48d43446-4a96-4385-8045-58ec84e7335f-logs\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.911631 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.911632 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-public-tls-certs\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.911674 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-config-data\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.911941 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:08 crc kubenswrapper[4735]: I0215 20:38:08.925931 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7n4j\" (UniqueName: \"kubernetes.io/projected/48d43446-4a96-4385-8045-58ec84e7335f-kube-api-access-f7n4j\") pod \"nova-api-0\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " pod="openstack/nova-api-0" Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.044870 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.295044 4735 generic.go:334] "Generic (PLEG): container finished" podID="f455dcec-5963-472c-9210-f111dac9c962" containerID="73405b46773b7734fa6b22df5d5b79a1a40dc2c9778063dbd1480450b0d29a92" exitCode=0 Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.295436 4735 generic.go:334] "Generic (PLEG): container finished" podID="f455dcec-5963-472c-9210-f111dac9c962" containerID="515f4b1cee2de2e7761314202a05e648f8ddae955a63237c7a7766c612abe2dd" exitCode=2 Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.295450 4735 generic.go:334] "Generic (PLEG): container finished" podID="f455dcec-5963-472c-9210-f111dac9c962" containerID="5d9ab85049e467b079c72760f19d754e9b1f1d542ef67118b0b6deba94769860" exitCode=0 Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.295892 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerDied","Data":"73405b46773b7734fa6b22df5d5b79a1a40dc2c9778063dbd1480450b0d29a92"} Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.315080 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerDied","Data":"515f4b1cee2de2e7761314202a05e648f8ddae955a63237c7a7766c612abe2dd"} Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.315093 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerDied","Data":"5d9ab85049e467b079c72760f19d754e9b1f1d542ef67118b0b6deba94769860"} Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.490743 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.589561 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.607111 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.608422 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 15 20:38:09 crc kubenswrapper[4735]: I0215 20:38:09.630247 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.306301 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"48d43446-4a96-4385-8045-58ec84e7335f","Type":"ContainerStarted","Data":"e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344"} Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.307415 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"48d43446-4a96-4385-8045-58ec84e7335f","Type":"ContainerStarted","Data":"49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef"} Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.307431 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"48d43446-4a96-4385-8045-58ec84e7335f","Type":"ContainerStarted","Data":"207c24bc68616a1f8b162d5962e9dc00e353c8a8c8844414161242a3d1457d7d"} Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.326340 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.3263219680000002 podStartE2EDuration="2.326321968s" podCreationTimestamp="2026-02-15 20:38:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:38:10.321239573 +0000 UTC m=+1298.187255196" watchObservedRunningTime="2026-02-15 20:38:10.326321968 +0000 UTC m=+1298.192337591" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.344460 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.506371 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-c647fbf5b-7zs9r" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.506410 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-9plzv"] Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.507683 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.508732 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.513497 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.515891 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9plzv"] Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.561830 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.590909 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d466d479d-wdxj6"] Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.630330 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.630551 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.674589 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-config-data\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.674699 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-scripts\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.674742 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.675339 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rvcb\" (UniqueName: \"kubernetes.io/projected/99051099-c063-4a04-a4a6-7f0e37e5c424-kube-api-access-8rvcb\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.777213 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-config-data\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.777301 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-scripts\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.777328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.777380 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rvcb\" (UniqueName: \"kubernetes.io/projected/99051099-c063-4a04-a4a6-7f0e37e5c424-kube-api-access-8rvcb\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.782472 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-scripts\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.782540 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.795614 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-config-data\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.802420 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rvcb\" (UniqueName: \"kubernetes.io/projected/99051099-c063-4a04-a4a6-7f0e37e5c424-kube-api-access-8rvcb\") pod \"nova-cell1-cell-mapping-9plzv\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:10 crc kubenswrapper[4735]: I0215 20:38:10.877094 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:11 crc kubenswrapper[4735]: I0215 20:38:11.317440 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon-log" containerID="cri-o://5d80cb6d2dc2ccf6ac05974c0b49e10cfa35444bc235b804bde522b9f6bb18fb" gracePeriod=30 Feb 15 20:38:11 crc kubenswrapper[4735]: I0215 20:38:11.317908 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" containerID="cri-o://67b53f3ca82355aac170c86227c023b886165282b3769c53e9f598b06b47689a" gracePeriod=30 Feb 15 20:38:11 crc kubenswrapper[4735]: I0215 20:38:11.412497 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9plzv"] Feb 15 20:38:11 crc kubenswrapper[4735]: W0215 20:38:11.414532 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99051099_c063_4a04_a4a6_7f0e37e5c424.slice/crio-5f2b6586bd1be2e9233362cad72e91dcc6afd6c1a877cb7fd6f3209241382bbd WatchSource:0}: Error finding container 5f2b6586bd1be2e9233362cad72e91dcc6afd6c1a877cb7fd6f3209241382bbd: Status 404 returned error can't find the container with id 5f2b6586bd1be2e9233362cad72e91dcc6afd6c1a877cb7fd6f3209241382bbd Feb 15 20:38:12 crc kubenswrapper[4735]: I0215 20:38:12.325682 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9plzv" event={"ID":"99051099-c063-4a04-a4a6-7f0e37e5c424","Type":"ContainerStarted","Data":"e333e9e509fcf421e4db4b2a9404903962b885dff8678be3b722771fc90fd4ed"} Feb 15 20:38:12 crc kubenswrapper[4735]: I0215 20:38:12.326043 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9plzv" event={"ID":"99051099-c063-4a04-a4a6-7f0e37e5c424","Type":"ContainerStarted","Data":"5f2b6586bd1be2e9233362cad72e91dcc6afd6c1a877cb7fd6f3209241382bbd"} Feb 15 20:38:12 crc kubenswrapper[4735]: I0215 20:38:12.352978 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-9plzv" podStartSLOduration=2.352958581 podStartE2EDuration="2.352958581s" podCreationTimestamp="2026-02-15 20:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:38:12.344083065 +0000 UTC m=+1300.210098688" watchObservedRunningTime="2026-02-15 20:38:12.352958581 +0000 UTC m=+1300.218974214" Feb 15 20:38:12 crc kubenswrapper[4735]: I0215 20:38:12.387198 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:38:12 crc kubenswrapper[4735]: I0215 20:38:12.450039 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-k7hsr"] Feb 15 20:38:12 crc kubenswrapper[4735]: I0215 20:38:12.450255 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" podUID="768dae67-d800-401b-b3d8-37438bc718ae" containerName="dnsmasq-dns" containerID="cri-o://821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2" gracePeriod=10 Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.049211 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.223392 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-nb\") pod \"768dae67-d800-401b-b3d8-37438bc718ae\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.223622 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-swift-storage-0\") pod \"768dae67-d800-401b-b3d8-37438bc718ae\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.223678 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-config\") pod \"768dae67-d800-401b-b3d8-37438bc718ae\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.223720 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktqbh\" (UniqueName: \"kubernetes.io/projected/768dae67-d800-401b-b3d8-37438bc718ae-kube-api-access-ktqbh\") pod \"768dae67-d800-401b-b3d8-37438bc718ae\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.223767 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-svc\") pod \"768dae67-d800-401b-b3d8-37438bc718ae\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.224298 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-sb\") pod \"768dae67-d800-401b-b3d8-37438bc718ae\" (UID: \"768dae67-d800-401b-b3d8-37438bc718ae\") " Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.232186 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/768dae67-d800-401b-b3d8-37438bc718ae-kube-api-access-ktqbh" (OuterVolumeSpecName: "kube-api-access-ktqbh") pod "768dae67-d800-401b-b3d8-37438bc718ae" (UID: "768dae67-d800-401b-b3d8-37438bc718ae"). InnerVolumeSpecName "kube-api-access-ktqbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.280846 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "768dae67-d800-401b-b3d8-37438bc718ae" (UID: "768dae67-d800-401b-b3d8-37438bc718ae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.311668 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-config" (OuterVolumeSpecName: "config") pod "768dae67-d800-401b-b3d8-37438bc718ae" (UID: "768dae67-d800-401b-b3d8-37438bc718ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.318311 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "768dae67-d800-401b-b3d8-37438bc718ae" (UID: "768dae67-d800-401b-b3d8-37438bc718ae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.324251 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "768dae67-d800-401b-b3d8-37438bc718ae" (UID: "768dae67-d800-401b-b3d8-37438bc718ae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.328878 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.328903 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktqbh\" (UniqueName: \"kubernetes.io/projected/768dae67-d800-401b-b3d8-37438bc718ae-kube-api-access-ktqbh\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.328913 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.328940 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.328959 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.333268 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "768dae67-d800-401b-b3d8-37438bc718ae" (UID: "768dae67-d800-401b-b3d8-37438bc718ae"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.343824 4735 generic.go:334] "Generic (PLEG): container finished" podID="768dae67-d800-401b-b3d8-37438bc718ae" containerID="821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2" exitCode=0 Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.344060 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.344100 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" event={"ID":"768dae67-d800-401b-b3d8-37438bc718ae","Type":"ContainerDied","Data":"821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2"} Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.344225 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-k7hsr" event={"ID":"768dae67-d800-401b-b3d8-37438bc718ae","Type":"ContainerDied","Data":"888af0908525aae5beeca7e3b620784497cb0fdbbdcaabb88076efc3277655c7"} Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.344253 4735 scope.go:117] "RemoveContainer" containerID="821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.370091 4735 scope.go:117] "RemoveContainer" containerID="eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.381209 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-k7hsr"] Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.389580 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-k7hsr"] Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.396395 4735 scope.go:117] "RemoveContainer" containerID="821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2" Feb 15 20:38:13 crc kubenswrapper[4735]: E0215 20:38:13.396878 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2\": container with ID starting with 821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2 not found: ID does not exist" containerID="821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.397008 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2"} err="failed to get container status \"821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2\": rpc error: code = NotFound desc = could not find container \"821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2\": container with ID starting with 821eb1f1f15fd9147e7b0a150a2886556acfaf3ad0480be9dfdbbd86df18aab2 not found: ID does not exist" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.397085 4735 scope.go:117] "RemoveContainer" containerID="eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96" Feb 15 20:38:13 crc kubenswrapper[4735]: E0215 20:38:13.397441 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96\": container with ID starting with eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96 not found: ID does not exist" containerID="eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.397480 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96"} err="failed to get container status \"eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96\": rpc error: code = NotFound desc = could not find container \"eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96\": container with ID starting with eaa3447e2b889e5068cc7b901479b80fa52ecf7b8f725cd93544a1f4778edf96 not found: ID does not exist" Feb 15 20:38:13 crc kubenswrapper[4735]: I0215 20:38:13.430788 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/768dae67-d800-401b-b3d8-37438bc718ae-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.359453 4735 generic.go:334] "Generic (PLEG): container finished" podID="f455dcec-5963-472c-9210-f111dac9c962" containerID="bb17c94ca3be19f8abbd9ca1803411e503157fd7a17c2c3d5c5f9d2b32feae1e" exitCode=0 Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.359733 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerDied","Data":"bb17c94ca3be19f8abbd9ca1803411e503157fd7a17c2c3d5c5f9d2b32feae1e"} Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.767407 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.778422 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.866335 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgbdl\" (UniqueName: \"kubernetes.io/projected/f455dcec-5963-472c-9210-f111dac9c962-kube-api-access-sgbdl\") pod \"f455dcec-5963-472c-9210-f111dac9c962\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.866422 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-ceilometer-tls-certs\") pod \"f455dcec-5963-472c-9210-f111dac9c962\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.866456 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-config-data\") pod \"f455dcec-5963-472c-9210-f111dac9c962\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.866476 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-run-httpd\") pod \"f455dcec-5963-472c-9210-f111dac9c962\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.866522 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-scripts\") pod \"f455dcec-5963-472c-9210-f111dac9c962\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.866641 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-sg-core-conf-yaml\") pod \"f455dcec-5963-472c-9210-f111dac9c962\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.866723 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-combined-ca-bundle\") pod \"f455dcec-5963-472c-9210-f111dac9c962\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.866755 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-log-httpd\") pod \"f455dcec-5963-472c-9210-f111dac9c962\" (UID: \"f455dcec-5963-472c-9210-f111dac9c962\") " Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.867489 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f455dcec-5963-472c-9210-f111dac9c962" (UID: "f455dcec-5963-472c-9210-f111dac9c962"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.867590 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f455dcec-5963-472c-9210-f111dac9c962" (UID: "f455dcec-5963-472c-9210-f111dac9c962"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.883100 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-scripts" (OuterVolumeSpecName: "scripts") pod "f455dcec-5963-472c-9210-f111dac9c962" (UID: "f455dcec-5963-472c-9210-f111dac9c962"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.891813 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f455dcec-5963-472c-9210-f111dac9c962-kube-api-access-sgbdl" (OuterVolumeSpecName: "kube-api-access-sgbdl") pod "f455dcec-5963-472c-9210-f111dac9c962" (UID: "f455dcec-5963-472c-9210-f111dac9c962"). InnerVolumeSpecName "kube-api-access-sgbdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.897023 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f455dcec-5963-472c-9210-f111dac9c962" (UID: "f455dcec-5963-472c-9210-f111dac9c962"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.904613 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="768dae67-d800-401b-b3d8-37438bc718ae" path="/var/lib/kubelet/pods/768dae67-d800-401b-b3d8-37438bc718ae/volumes" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.950397 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f455dcec-5963-472c-9210-f111dac9c962" (UID: "f455dcec-5963-472c-9210-f111dac9c962"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.969881 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.969905 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.969914 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgbdl\" (UniqueName: \"kubernetes.io/projected/f455dcec-5963-472c-9210-f111dac9c962-kube-api-access-sgbdl\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.969923 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.969931 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f455dcec-5963-472c-9210-f111dac9c962-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:14 crc kubenswrapper[4735]: I0215 20:38:14.970009 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.007050 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f455dcec-5963-472c-9210-f111dac9c962" (UID: "f455dcec-5963-472c-9210-f111dac9c962"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.015290 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-config-data" (OuterVolumeSpecName: "config-data") pod "f455dcec-5963-472c-9210-f111dac9c962" (UID: "f455dcec-5963-472c-9210-f111dac9c962"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.072806 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.072840 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f455dcec-5963-472c-9210-f111dac9c962-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.372115 4735 generic.go:334] "Generic (PLEG): container finished" podID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerID="67b53f3ca82355aac170c86227c023b886165282b3769c53e9f598b06b47689a" exitCode=0 Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.372210 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerDied","Data":"67b53f3ca82355aac170c86227c023b886165282b3769c53e9f598b06b47689a"} Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.372412 4735 scope.go:117] "RemoveContainer" containerID="86ffd054acc51bbf8b9d87f36ab9440eeb9c687baf017ae70a357fa95727f187" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.378600 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f455dcec-5963-472c-9210-f111dac9c962","Type":"ContainerDied","Data":"b7f1e5e84c2be750f1d5bc6be76b34812120db7898bc186aaf08bd8a489c226c"} Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.378933 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.454365 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.488037 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.495996 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:15 crc kubenswrapper[4735]: E0215 20:38:15.496491 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="proxy-httpd" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.496517 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="proxy-httpd" Feb 15 20:38:15 crc kubenswrapper[4735]: E0215 20:38:15.496536 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="768dae67-d800-401b-b3d8-37438bc718ae" containerName="init" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.496545 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="768dae67-d800-401b-b3d8-37438bc718ae" containerName="init" Feb 15 20:38:15 crc kubenswrapper[4735]: E0215 20:38:15.496560 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="sg-core" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.496567 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="sg-core" Feb 15 20:38:15 crc kubenswrapper[4735]: E0215 20:38:15.496579 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="ceilometer-central-agent" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.496586 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="ceilometer-central-agent" Feb 15 20:38:15 crc kubenswrapper[4735]: E0215 20:38:15.496602 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="ceilometer-notification-agent" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.496610 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="ceilometer-notification-agent" Feb 15 20:38:15 crc kubenswrapper[4735]: E0215 20:38:15.496628 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="768dae67-d800-401b-b3d8-37438bc718ae" containerName="dnsmasq-dns" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.496635 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="768dae67-d800-401b-b3d8-37438bc718ae" containerName="dnsmasq-dns" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.497229 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="ceilometer-notification-agent" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.497258 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="proxy-httpd" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.497273 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="768dae67-d800-401b-b3d8-37438bc718ae" containerName="dnsmasq-dns" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.497291 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="ceilometer-central-agent" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.497304 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f455dcec-5963-472c-9210-f111dac9c962" containerName="sg-core" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.499559 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.503155 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.505037 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.505039 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.506809 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.563116 4735 scope.go:117] "RemoveContainer" containerID="73405b46773b7734fa6b22df5d5b79a1a40dc2c9778063dbd1480450b0d29a92" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.590133 4735 scope.go:117] "RemoveContainer" containerID="515f4b1cee2de2e7761314202a05e648f8ddae955a63237c7a7766c612abe2dd" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.616193 4735 scope.go:117] "RemoveContainer" containerID="5d9ab85049e467b079c72760f19d754e9b1f1d542ef67118b0b6deba94769860" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.635248 4735 scope.go:117] "RemoveContainer" containerID="bb17c94ca3be19f8abbd9ca1803411e503157fd7a17c2c3d5c5f9d2b32feae1e" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.685617 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-scripts\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.685680 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tcsz\" (UniqueName: \"kubernetes.io/projected/09208797-cc26-4216-b27f-51b281eba74a-kube-api-access-8tcsz\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.685779 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09208797-cc26-4216-b27f-51b281eba74a-run-httpd\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.685811 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-config-data\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.685869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.685919 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09208797-cc26-4216-b27f-51b281eba74a-log-httpd\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.686000 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.686028 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.787436 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.787511 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09208797-cc26-4216-b27f-51b281eba74a-log-httpd\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.787563 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.787585 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.787613 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-scripts\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.787634 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tcsz\" (UniqueName: \"kubernetes.io/projected/09208797-cc26-4216-b27f-51b281eba74a-kube-api-access-8tcsz\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.787677 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09208797-cc26-4216-b27f-51b281eba74a-run-httpd\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.787699 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-config-data\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.788510 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09208797-cc26-4216-b27f-51b281eba74a-run-httpd\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.789819 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09208797-cc26-4216-b27f-51b281eba74a-log-httpd\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.795918 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-config-data\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.796665 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.797084 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.799012 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-scripts\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.804507 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/09208797-cc26-4216-b27f-51b281eba74a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.807775 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tcsz\" (UniqueName: \"kubernetes.io/projected/09208797-cc26-4216-b27f-51b281eba74a-kube-api-access-8tcsz\") pod \"ceilometer-0\" (UID: \"09208797-cc26-4216-b27f-51b281eba74a\") " pod="openstack/ceilometer-0" Feb 15 20:38:15 crc kubenswrapper[4735]: I0215 20:38:15.822921 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 15 20:38:16 crc kubenswrapper[4735]: I0215 20:38:16.544460 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 15 20:38:16 crc kubenswrapper[4735]: W0215 20:38:16.547763 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09208797_cc26_4216_b27f_51b281eba74a.slice/crio-5d856e5f7cbac079faf2f3e8f0ea07e14b556ee37c405e483182841b1a53bbd1 WatchSource:0}: Error finding container 5d856e5f7cbac079faf2f3e8f0ea07e14b556ee37c405e483182841b1a53bbd1: Status 404 returned error can't find the container with id 5d856e5f7cbac079faf2f3e8f0ea07e14b556ee37c405e483182841b1a53bbd1 Feb 15 20:38:16 crc kubenswrapper[4735]: I0215 20:38:16.907067 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f455dcec-5963-472c-9210-f111dac9c962" path="/var/lib/kubelet/pods/f455dcec-5963-472c-9210-f111dac9c962/volumes" Feb 15 20:38:17 crc kubenswrapper[4735]: I0215 20:38:17.405535 4735 generic.go:334] "Generic (PLEG): container finished" podID="99051099-c063-4a04-a4a6-7f0e37e5c424" containerID="e333e9e509fcf421e4db4b2a9404903962b885dff8678be3b722771fc90fd4ed" exitCode=0 Feb 15 20:38:17 crc kubenswrapper[4735]: I0215 20:38:17.405649 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9plzv" event={"ID":"99051099-c063-4a04-a4a6-7f0e37e5c424","Type":"ContainerDied","Data":"e333e9e509fcf421e4db4b2a9404903962b885dff8678be3b722771fc90fd4ed"} Feb 15 20:38:17 crc kubenswrapper[4735]: I0215 20:38:17.411482 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09208797-cc26-4216-b27f-51b281eba74a","Type":"ContainerStarted","Data":"65ed6983e27772d9caef0c113ed0e5741b9ca7dd8a49510c162ea4e37dae9af1"} Feb 15 20:38:17 crc kubenswrapper[4735]: I0215 20:38:17.411526 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09208797-cc26-4216-b27f-51b281eba74a","Type":"ContainerStarted","Data":"5d856e5f7cbac079faf2f3e8f0ea07e14b556ee37c405e483182841b1a53bbd1"} Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.425464 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09208797-cc26-4216-b27f-51b281eba74a","Type":"ContainerStarted","Data":"00f1860eac2927c606096f6b3e6f16658eda59c1558125a156b5d62ecdf1ae22"} Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.738855 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.909777 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rvcb\" (UniqueName: \"kubernetes.io/projected/99051099-c063-4a04-a4a6-7f0e37e5c424-kube-api-access-8rvcb\") pod \"99051099-c063-4a04-a4a6-7f0e37e5c424\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.909936 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-config-data\") pod \"99051099-c063-4a04-a4a6-7f0e37e5c424\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.909985 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-scripts\") pod \"99051099-c063-4a04-a4a6-7f0e37e5c424\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.910033 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-combined-ca-bundle\") pod \"99051099-c063-4a04-a4a6-7f0e37e5c424\" (UID: \"99051099-c063-4a04-a4a6-7f0e37e5c424\") " Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.916387 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-scripts" (OuterVolumeSpecName: "scripts") pod "99051099-c063-4a04-a4a6-7f0e37e5c424" (UID: "99051099-c063-4a04-a4a6-7f0e37e5c424"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.917179 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99051099-c063-4a04-a4a6-7f0e37e5c424-kube-api-access-8rvcb" (OuterVolumeSpecName: "kube-api-access-8rvcb") pod "99051099-c063-4a04-a4a6-7f0e37e5c424" (UID: "99051099-c063-4a04-a4a6-7f0e37e5c424"). InnerVolumeSpecName "kube-api-access-8rvcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.938747 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-config-data" (OuterVolumeSpecName: "config-data") pod "99051099-c063-4a04-a4a6-7f0e37e5c424" (UID: "99051099-c063-4a04-a4a6-7f0e37e5c424"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:18 crc kubenswrapper[4735]: I0215 20:38:18.950366 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99051099-c063-4a04-a4a6-7f0e37e5c424" (UID: "99051099-c063-4a04-a4a6-7f0e37e5c424"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.011678 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rvcb\" (UniqueName: \"kubernetes.io/projected/99051099-c063-4a04-a4a6-7f0e37e5c424-kube-api-access-8rvcb\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.011712 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.011722 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.011734 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99051099-c063-4a04-a4a6-7f0e37e5c424-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.045919 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.047503 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.434923 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9plzv" event={"ID":"99051099-c063-4a04-a4a6-7f0e37e5c424","Type":"ContainerDied","Data":"5f2b6586bd1be2e9233362cad72e91dcc6afd6c1a877cb7fd6f3209241382bbd"} Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.436027 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f2b6586bd1be2e9233362cad72e91dcc6afd6c1a877cb7fd6f3209241382bbd" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.434967 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9plzv" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.437510 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09208797-cc26-4216-b27f-51b281eba74a","Type":"ContainerStarted","Data":"6ce9457068d532049c96f2f2be77f43f8d413487ee081075813814257d013cff"} Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.614213 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.651773 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.657326 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.663406 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.664087 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.664490 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="36959d4f-b1ae-491c-a24c-a437aa21ff06" containerName="nova-scheduler-scheduler" containerID="cri-o://794872a580ea452d0045b46c12bad3839f4202d8b7514046171f4f99a8403864" gracePeriod=30 Feb 15 20:38:19 crc kubenswrapper[4735]: I0215 20:38:19.731929 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:38:20 crc kubenswrapper[4735]: I0215 20:38:20.059104 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 15 20:38:20 crc kubenswrapper[4735]: I0215 20:38:20.059129 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 15 20:38:20 crc kubenswrapper[4735]: I0215 20:38:20.466689 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-log" containerID="cri-o://49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef" gracePeriod=30 Feb 15 20:38:20 crc kubenswrapper[4735]: I0215 20:38:20.470597 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09208797-cc26-4216-b27f-51b281eba74a","Type":"ContainerStarted","Data":"b4d1277697f83d20ccacc672edad367f94ea0193dc83d70513b9fdcf963d22bc"} Feb 15 20:38:20 crc kubenswrapper[4735]: I0215 20:38:20.470639 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 15 20:38:20 crc kubenswrapper[4735]: I0215 20:38:20.470914 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-api" containerID="cri-o://e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344" gracePeriod=30 Feb 15 20:38:20 crc kubenswrapper[4735]: I0215 20:38:20.521750 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 15 20:38:20 crc kubenswrapper[4735]: I0215 20:38:20.525921 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.203198159 podStartE2EDuration="5.525904934s" podCreationTimestamp="2026-02-15 20:38:15 +0000 UTC" firstStartedPulling="2026-02-15 20:38:16.550172035 +0000 UTC m=+1304.416187658" lastFinishedPulling="2026-02-15 20:38:19.87287881 +0000 UTC m=+1307.738894433" observedRunningTime="2026-02-15 20:38:20.506766584 +0000 UTC m=+1308.372782207" watchObservedRunningTime="2026-02-15 20:38:20.525904934 +0000 UTC m=+1308.391920557" Feb 15 20:38:21 crc kubenswrapper[4735]: E0215 20:38:21.369176 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="794872a580ea452d0045b46c12bad3839f4202d8b7514046171f4f99a8403864" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 15 20:38:21 crc kubenswrapper[4735]: E0215 20:38:21.370986 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="794872a580ea452d0045b46c12bad3839f4202d8b7514046171f4f99a8403864" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 15 20:38:21 crc kubenswrapper[4735]: E0215 20:38:21.372550 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="794872a580ea452d0045b46c12bad3839f4202d8b7514046171f4f99a8403864" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 15 20:38:21 crc kubenswrapper[4735]: E0215 20:38:21.372616 4735 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="36959d4f-b1ae-491c-a24c-a437aa21ff06" containerName="nova-scheduler-scheduler" Feb 15 20:38:21 crc kubenswrapper[4735]: I0215 20:38:21.473646 4735 generic.go:334] "Generic (PLEG): container finished" podID="48d43446-4a96-4385-8045-58ec84e7335f" containerID="49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef" exitCode=143 Feb 15 20:38:21 crc kubenswrapper[4735]: I0215 20:38:21.473710 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"48d43446-4a96-4385-8045-58ec84e7335f","Type":"ContainerDied","Data":"49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef"} Feb 15 20:38:21 crc kubenswrapper[4735]: I0215 20:38:21.474220 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-log" containerID="cri-o://637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71" gracePeriod=30 Feb 15 20:38:21 crc kubenswrapper[4735]: I0215 20:38:21.474477 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-metadata" containerID="cri-o://e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85" gracePeriod=30 Feb 15 20:38:22 crc kubenswrapper[4735]: I0215 20:38:22.489627 4735 generic.go:334] "Generic (PLEG): container finished" podID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerID="637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71" exitCode=143 Feb 15 20:38:22 crc kubenswrapper[4735]: I0215 20:38:22.489676 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fccd554-a077-4c87-931f-5c1c6040ec9e","Type":"ContainerDied","Data":"637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71"} Feb 15 20:38:24 crc kubenswrapper[4735]: I0215 20:38:24.613625 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:41994->10.217.0.204:8775: read: connection reset by peer" Feb 15 20:38:24 crc kubenswrapper[4735]: I0215 20:38:24.613675 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:41998->10.217.0.204:8775: read: connection reset by peer" Feb 15 20:38:24 crc kubenswrapper[4735]: I0215 20:38:24.781729 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.082378 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.137718 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-combined-ca-bundle\") pod \"7fccd554-a077-4c87-931f-5c1c6040ec9e\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.137831 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqmc5\" (UniqueName: \"kubernetes.io/projected/7fccd554-a077-4c87-931f-5c1c6040ec9e-kube-api-access-hqmc5\") pod \"7fccd554-a077-4c87-931f-5c1c6040ec9e\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.137891 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fccd554-a077-4c87-931f-5c1c6040ec9e-logs\") pod \"7fccd554-a077-4c87-931f-5c1c6040ec9e\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.137931 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-nova-metadata-tls-certs\") pod \"7fccd554-a077-4c87-931f-5c1c6040ec9e\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.137984 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-config-data\") pod \"7fccd554-a077-4c87-931f-5c1c6040ec9e\" (UID: \"7fccd554-a077-4c87-931f-5c1c6040ec9e\") " Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.141593 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fccd554-a077-4c87-931f-5c1c6040ec9e-logs" (OuterVolumeSpecName: "logs") pod "7fccd554-a077-4c87-931f-5c1c6040ec9e" (UID: "7fccd554-a077-4c87-931f-5c1c6040ec9e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.169338 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fccd554-a077-4c87-931f-5c1c6040ec9e-kube-api-access-hqmc5" (OuterVolumeSpecName: "kube-api-access-hqmc5") pod "7fccd554-a077-4c87-931f-5c1c6040ec9e" (UID: "7fccd554-a077-4c87-931f-5c1c6040ec9e"). InnerVolumeSpecName "kube-api-access-hqmc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.206190 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-config-data" (OuterVolumeSpecName: "config-data") pod "7fccd554-a077-4c87-931f-5c1c6040ec9e" (UID: "7fccd554-a077-4c87-931f-5c1c6040ec9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.240783 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqmc5\" (UniqueName: \"kubernetes.io/projected/7fccd554-a077-4c87-931f-5c1c6040ec9e-kube-api-access-hqmc5\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.241110 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fccd554-a077-4c87-931f-5c1c6040ec9e-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.241215 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.241054 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fccd554-a077-4c87-931f-5c1c6040ec9e" (UID: "7fccd554-a077-4c87-931f-5c1c6040ec9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.255753 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7fccd554-a077-4c87-931f-5c1c6040ec9e" (UID: "7fccd554-a077-4c87-931f-5c1c6040ec9e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.343759 4735 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.344079 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fccd554-a077-4c87-931f-5c1c6040ec9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.527373 4735 generic.go:334] "Generic (PLEG): container finished" podID="36959d4f-b1ae-491c-a24c-a437aa21ff06" containerID="794872a580ea452d0045b46c12bad3839f4202d8b7514046171f4f99a8403864" exitCode=0 Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.527433 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"36959d4f-b1ae-491c-a24c-a437aa21ff06","Type":"ContainerDied","Data":"794872a580ea452d0045b46c12bad3839f4202d8b7514046171f4f99a8403864"} Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.528934 4735 generic.go:334] "Generic (PLEG): container finished" podID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerID="e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85" exitCode=0 Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.528968 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fccd554-a077-4c87-931f-5c1c6040ec9e","Type":"ContainerDied","Data":"e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85"} Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.528985 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fccd554-a077-4c87-931f-5c1c6040ec9e","Type":"ContainerDied","Data":"5fa80fc3609f09fdf5152c78d8f9672d570f7052abc555c787f6a60ff39b025e"} Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.529002 4735 scope.go:117] "RemoveContainer" containerID="e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.529092 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.532493 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.564347 4735 scope.go:117] "RemoveContainer" containerID="637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.579707 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.606749 4735 scope.go:117] "RemoveContainer" containerID="e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85" Feb 15 20:38:25 crc kubenswrapper[4735]: E0215 20:38:25.607213 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85\": container with ID starting with e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85 not found: ID does not exist" containerID="e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.607246 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85"} err="failed to get container status \"e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85\": rpc error: code = NotFound desc = could not find container \"e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85\": container with ID starting with e8d3d630b3f8440e287fd689be41435261b0f00e8ea3a5eb0f68f93f26430e85 not found: ID does not exist" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.607266 4735 scope.go:117] "RemoveContainer" containerID="637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71" Feb 15 20:38:25 crc kubenswrapper[4735]: E0215 20:38:25.607533 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71\": container with ID starting with 637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71 not found: ID does not exist" containerID="637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.607554 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71"} err="failed to get container status \"637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71\": rpc error: code = NotFound desc = could not find container \"637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71\": container with ID starting with 637e4b310b0e0fbf61d5a118dd5e1c0e1d3dbd4491b5d31fc1ff1f560b2dba71 not found: ID does not exist" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.614290 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.620216 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:38:25 crc kubenswrapper[4735]: E0215 20:38:25.620563 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36959d4f-b1ae-491c-a24c-a437aa21ff06" containerName="nova-scheduler-scheduler" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.620580 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="36959d4f-b1ae-491c-a24c-a437aa21ff06" containerName="nova-scheduler-scheduler" Feb 15 20:38:25 crc kubenswrapper[4735]: E0215 20:38:25.620609 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99051099-c063-4a04-a4a6-7f0e37e5c424" containerName="nova-manage" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.620616 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="99051099-c063-4a04-a4a6-7f0e37e5c424" containerName="nova-manage" Feb 15 20:38:25 crc kubenswrapper[4735]: E0215 20:38:25.620627 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-metadata" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.620634 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-metadata" Feb 15 20:38:25 crc kubenswrapper[4735]: E0215 20:38:25.620655 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-log" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.620660 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-log" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.620823 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-log" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.620841 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" containerName="nova-metadata-metadata" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.620853 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="99051099-c063-4a04-a4a6-7f0e37e5c424" containerName="nova-manage" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.620862 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="36959d4f-b1ae-491c-a24c-a437aa21ff06" containerName="nova-scheduler-scheduler" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.621822 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.625186 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.625340 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.634345 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.649542 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-combined-ca-bundle\") pod \"36959d4f-b1ae-491c-a24c-a437aa21ff06\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.649712 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wh8h\" (UniqueName: \"kubernetes.io/projected/36959d4f-b1ae-491c-a24c-a437aa21ff06-kube-api-access-7wh8h\") pod \"36959d4f-b1ae-491c-a24c-a437aa21ff06\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.649819 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-config-data\") pod \"36959d4f-b1ae-491c-a24c-a437aa21ff06\" (UID: \"36959d4f-b1ae-491c-a24c-a437aa21ff06\") " Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.650088 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qpbp\" (UniqueName: \"kubernetes.io/projected/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-kube-api-access-5qpbp\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.650317 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-logs\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.650490 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.650542 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.650605 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-config-data\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.656159 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36959d4f-b1ae-491c-a24c-a437aa21ff06-kube-api-access-7wh8h" (OuterVolumeSpecName: "kube-api-access-7wh8h") pod "36959d4f-b1ae-491c-a24c-a437aa21ff06" (UID: "36959d4f-b1ae-491c-a24c-a437aa21ff06"). InnerVolumeSpecName "kube-api-access-7wh8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.680841 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-config-data" (OuterVolumeSpecName: "config-data") pod "36959d4f-b1ae-491c-a24c-a437aa21ff06" (UID: "36959d4f-b1ae-491c-a24c-a437aa21ff06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.693472 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36959d4f-b1ae-491c-a24c-a437aa21ff06" (UID: "36959d4f-b1ae-491c-a24c-a437aa21ff06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.751900 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-logs\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.752004 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.752036 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.752067 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-config-data\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.752119 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qpbp\" (UniqueName: \"kubernetes.io/projected/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-kube-api-access-5qpbp\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.752172 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wh8h\" (UniqueName: \"kubernetes.io/projected/36959d4f-b1ae-491c-a24c-a437aa21ff06-kube-api-access-7wh8h\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.752186 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.752195 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36959d4f-b1ae-491c-a24c-a437aa21ff06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.753610 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-logs\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.755595 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.756299 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-config-data\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.757616 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.773833 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qpbp\" (UniqueName: \"kubernetes.io/projected/90c9c199-a27c-464a-8845-c2fd4e0b0fd9-kube-api-access-5qpbp\") pod \"nova-metadata-0\" (UID: \"90c9c199-a27c-464a-8845-c2fd4e0b0fd9\") " pod="openstack/nova-metadata-0" Feb 15 20:38:25 crc kubenswrapper[4735]: I0215 20:38:25.966718 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.314073 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.401044 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7n4j\" (UniqueName: \"kubernetes.io/projected/48d43446-4a96-4385-8045-58ec84e7335f-kube-api-access-f7n4j\") pod \"48d43446-4a96-4385-8045-58ec84e7335f\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.401167 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-combined-ca-bundle\") pod \"48d43446-4a96-4385-8045-58ec84e7335f\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.401225 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-internal-tls-certs\") pod \"48d43446-4a96-4385-8045-58ec84e7335f\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.401280 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-config-data\") pod \"48d43446-4a96-4385-8045-58ec84e7335f\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.401340 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48d43446-4a96-4385-8045-58ec84e7335f-logs\") pod \"48d43446-4a96-4385-8045-58ec84e7335f\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.401485 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-public-tls-certs\") pod \"48d43446-4a96-4385-8045-58ec84e7335f\" (UID: \"48d43446-4a96-4385-8045-58ec84e7335f\") " Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.402163 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48d43446-4a96-4385-8045-58ec84e7335f-logs" (OuterVolumeSpecName: "logs") pod "48d43446-4a96-4385-8045-58ec84e7335f" (UID: "48d43446-4a96-4385-8045-58ec84e7335f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.407515 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d43446-4a96-4385-8045-58ec84e7335f-kube-api-access-f7n4j" (OuterVolumeSpecName: "kube-api-access-f7n4j") pod "48d43446-4a96-4385-8045-58ec84e7335f" (UID: "48d43446-4a96-4385-8045-58ec84e7335f"). InnerVolumeSpecName "kube-api-access-f7n4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.437209 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48d43446-4a96-4385-8045-58ec84e7335f" (UID: "48d43446-4a96-4385-8045-58ec84e7335f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.445142 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-config-data" (OuterVolumeSpecName: "config-data") pod "48d43446-4a96-4385-8045-58ec84e7335f" (UID: "48d43446-4a96-4385-8045-58ec84e7335f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.461652 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "48d43446-4a96-4385-8045-58ec84e7335f" (UID: "48d43446-4a96-4385-8045-58ec84e7335f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.475751 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "48d43446-4a96-4385-8045-58ec84e7335f" (UID: "48d43446-4a96-4385-8045-58ec84e7335f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.504694 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.505000 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7n4j\" (UniqueName: \"kubernetes.io/projected/48d43446-4a96-4385-8045-58ec84e7335f-kube-api-access-f7n4j\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.505093 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.505163 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.505228 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48d43446-4a96-4385-8045-58ec84e7335f-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.505283 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48d43446-4a96-4385-8045-58ec84e7335f-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.538254 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"36959d4f-b1ae-491c-a24c-a437aa21ff06","Type":"ContainerDied","Data":"9c4fe031e649cbb773ac4bb6611853d6d8523928a9bd08d277f907ac0495f46f"} Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.538311 4735 scope.go:117] "RemoveContainer" containerID="794872a580ea452d0045b46c12bad3839f4202d8b7514046171f4f99a8403864" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.538429 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.541558 4735 generic.go:334] "Generic (PLEG): container finished" podID="48d43446-4a96-4385-8045-58ec84e7335f" containerID="e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344" exitCode=0 Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.541639 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"48d43446-4a96-4385-8045-58ec84e7335f","Type":"ContainerDied","Data":"e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344"} Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.541670 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"48d43446-4a96-4385-8045-58ec84e7335f","Type":"ContainerDied","Data":"207c24bc68616a1f8b162d5962e9dc00e353c8a8c8844414161242a3d1457d7d"} Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.541726 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.570637 4735 scope.go:117] "RemoveContainer" containerID="e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.583640 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.591760 4735 scope.go:117] "RemoveContainer" containerID="49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.608127 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.622032 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.636758 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.662380 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.683154 4735 scope.go:117] "RemoveContainer" containerID="e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.683274 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:26 crc kubenswrapper[4735]: E0215 20:38:26.684117 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-log" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.698088 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-log" Feb 15 20:38:26 crc kubenswrapper[4735]: E0215 20:38:26.698474 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-api" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.698567 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-api" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.699396 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-api" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.699511 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="48d43446-4a96-4385-8045-58ec84e7335f" containerName="nova-api-log" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.701429 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: E0215 20:38:26.690455 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344\": container with ID starting with e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344 not found: ID does not exist" containerID="e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.707469 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344"} err="failed to get container status \"e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344\": rpc error: code = NotFound desc = could not find container \"e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344\": container with ID starting with e664a5e73d251df79df1cf5cb6a333ff65e012d7c910de6e03a899d806207344 not found: ID does not exist" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.707498 4735 scope.go:117] "RemoveContainer" containerID="49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef" Feb 15 20:38:26 crc kubenswrapper[4735]: E0215 20:38:26.708714 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef\": container with ID starting with 49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef not found: ID does not exist" containerID="49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.708820 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef"} err="failed to get container status \"49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef\": rpc error: code = NotFound desc = could not find container \"49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef\": container with ID starting with 49e79622efdfb6b72adc550638f9384015a26d7f2d92c548a137275f4e6eb5ef not found: ID does not exist" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.713622 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.714112 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.714445 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.742103 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.743868 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.758327 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.764505 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.792198 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.811036 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mpxw\" (UniqueName: \"kubernetes.io/projected/a964a23e-2012-4d43-98c2-e0b15f9d9d74-kube-api-access-2mpxw\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.811183 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2066599-479d-43c3-86cb-528fc3922810-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d2066599-479d-43c3-86cb-528fc3922810\") " pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.811247 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2066599-479d-43c3-86cb-528fc3922810-config-data\") pod \"nova-scheduler-0\" (UID: \"d2066599-479d-43c3-86cb-528fc3922810\") " pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.811267 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snkvj\" (UniqueName: \"kubernetes.io/projected/d2066599-479d-43c3-86cb-528fc3922810-kube-api-access-snkvj\") pod \"nova-scheduler-0\" (UID: \"d2066599-479d-43c3-86cb-528fc3922810\") " pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.811295 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.811345 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-config-data\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.811377 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a964a23e-2012-4d43-98c2-e0b15f9d9d74-logs\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.811409 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.811424 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-public-tls-certs\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.900097 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36959d4f-b1ae-491c-a24c-a437aa21ff06" path="/var/lib/kubelet/pods/36959d4f-b1ae-491c-a24c-a437aa21ff06/volumes" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.900867 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48d43446-4a96-4385-8045-58ec84e7335f" path="/var/lib/kubelet/pods/48d43446-4a96-4385-8045-58ec84e7335f/volumes" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.901606 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fccd554-a077-4c87-931f-5c1c6040ec9e" path="/var/lib/kubelet/pods/7fccd554-a077-4c87-931f-5c1c6040ec9e/volumes" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.912598 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2066599-479d-43c3-86cb-528fc3922810-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d2066599-479d-43c3-86cb-528fc3922810\") " pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.912753 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2066599-479d-43c3-86cb-528fc3922810-config-data\") pod \"nova-scheduler-0\" (UID: \"d2066599-479d-43c3-86cb-528fc3922810\") " pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.912835 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snkvj\" (UniqueName: \"kubernetes.io/projected/d2066599-479d-43c3-86cb-528fc3922810-kube-api-access-snkvj\") pod \"nova-scheduler-0\" (UID: \"d2066599-479d-43c3-86cb-528fc3922810\") " pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.912912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.913016 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-config-data\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.913105 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a964a23e-2012-4d43-98c2-e0b15f9d9d74-logs\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.913217 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.913293 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-public-tls-certs\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.913382 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mpxw\" (UniqueName: \"kubernetes.io/projected/a964a23e-2012-4d43-98c2-e0b15f9d9d74-kube-api-access-2mpxw\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.914554 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a964a23e-2012-4d43-98c2-e0b15f9d9d74-logs\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.917473 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2066599-479d-43c3-86cb-528fc3922810-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d2066599-479d-43c3-86cb-528fc3922810\") " pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.917885 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2066599-479d-43c3-86cb-528fc3922810-config-data\") pod \"nova-scheduler-0\" (UID: \"d2066599-479d-43c3-86cb-528fc3922810\") " pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.926557 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-config-data\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.926989 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.927864 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-public-tls-certs\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.928628 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a964a23e-2012-4d43-98c2-e0b15f9d9d74-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.930842 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snkvj\" (UniqueName: \"kubernetes.io/projected/d2066599-479d-43c3-86cb-528fc3922810-kube-api-access-snkvj\") pod \"nova-scheduler-0\" (UID: \"d2066599-479d-43c3-86cb-528fc3922810\") " pod="openstack/nova-scheduler-0" Feb 15 20:38:26 crc kubenswrapper[4735]: I0215 20:38:26.931134 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mpxw\" (UniqueName: \"kubernetes.io/projected/a964a23e-2012-4d43-98c2-e0b15f9d9d74-kube-api-access-2mpxw\") pod \"nova-api-0\" (UID: \"a964a23e-2012-4d43-98c2-e0b15f9d9d74\") " pod="openstack/nova-api-0" Feb 15 20:38:27 crc kubenswrapper[4735]: I0215 20:38:27.132103 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 15 20:38:27 crc kubenswrapper[4735]: I0215 20:38:27.160571 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 15 20:38:27 crc kubenswrapper[4735]: I0215 20:38:27.553357 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"90c9c199-a27c-464a-8845-c2fd4e0b0fd9","Type":"ContainerStarted","Data":"5c7289b509bb11943b2d6c97d4c8503771418f2d58852a56ecee7420214297a0"} Feb 15 20:38:27 crc kubenswrapper[4735]: I0215 20:38:27.553701 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"90c9c199-a27c-464a-8845-c2fd4e0b0fd9","Type":"ContainerStarted","Data":"579d866ee1bf6329f86a95377ebc8e4167991df1c0418509d9e5e36090247481"} Feb 15 20:38:27 crc kubenswrapper[4735]: I0215 20:38:27.553716 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"90c9c199-a27c-464a-8845-c2fd4e0b0fd9","Type":"ContainerStarted","Data":"c8682bfac3885f9293b6cf437be131c5fe0db2e9beabb423da80fc26e1443546"} Feb 15 20:38:27 crc kubenswrapper[4735]: I0215 20:38:27.576522 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.576502205 podStartE2EDuration="2.576502205s" podCreationTimestamp="2026-02-15 20:38:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:38:27.570976567 +0000 UTC m=+1315.436992190" watchObservedRunningTime="2026-02-15 20:38:27.576502205 +0000 UTC m=+1315.442517848" Feb 15 20:38:27 crc kubenswrapper[4735]: W0215 20:38:27.625914 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda964a23e_2012_4d43_98c2_e0b15f9d9d74.slice/crio-aca55d2797d55ce0dfaca8b37a6426e1f68503fe6b47af0a680116bbf6b125da WatchSource:0}: Error finding container aca55d2797d55ce0dfaca8b37a6426e1f68503fe6b47af0a680116bbf6b125da: Status 404 returned error can't find the container with id aca55d2797d55ce0dfaca8b37a6426e1f68503fe6b47af0a680116bbf6b125da Feb 15 20:38:27 crc kubenswrapper[4735]: I0215 20:38:27.629056 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 15 20:38:27 crc kubenswrapper[4735]: I0215 20:38:27.713185 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 15 20:38:27 crc kubenswrapper[4735]: W0215 20:38:27.713777 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2066599_479d_43c3_86cb_528fc3922810.slice/crio-30a9afe3e41281bfa510743246cb34f8977cbd17a167575288671463a88e5347 WatchSource:0}: Error finding container 30a9afe3e41281bfa510743246cb34f8977cbd17a167575288671463a88e5347: Status 404 returned error can't find the container with id 30a9afe3e41281bfa510743246cb34f8977cbd17a167575288671463a88e5347 Feb 15 20:38:28 crc kubenswrapper[4735]: I0215 20:38:28.567996 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a964a23e-2012-4d43-98c2-e0b15f9d9d74","Type":"ContainerStarted","Data":"b1c9ab8ded1e1e93be00a04c3605200e4e0d82a1cca767283529785c3e23c2a4"} Feb 15 20:38:28 crc kubenswrapper[4735]: I0215 20:38:28.568043 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a964a23e-2012-4d43-98c2-e0b15f9d9d74","Type":"ContainerStarted","Data":"0ece5df993dffacc936f6d1432a2296c5751d2adb71944429e1fd1236a4d450d"} Feb 15 20:38:28 crc kubenswrapper[4735]: I0215 20:38:28.568056 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a964a23e-2012-4d43-98c2-e0b15f9d9d74","Type":"ContainerStarted","Data":"aca55d2797d55ce0dfaca8b37a6426e1f68503fe6b47af0a680116bbf6b125da"} Feb 15 20:38:28 crc kubenswrapper[4735]: I0215 20:38:28.570239 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d2066599-479d-43c3-86cb-528fc3922810","Type":"ContainerStarted","Data":"5ba254c8175fde4586afdbb58b3002dd37b94811d9e82091bd5e06267673dce7"} Feb 15 20:38:28 crc kubenswrapper[4735]: I0215 20:38:28.570491 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d2066599-479d-43c3-86cb-528fc3922810","Type":"ContainerStarted","Data":"30a9afe3e41281bfa510743246cb34f8977cbd17a167575288671463a88e5347"} Feb 15 20:38:28 crc kubenswrapper[4735]: I0215 20:38:28.594496 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.594476965 podStartE2EDuration="2.594476965s" podCreationTimestamp="2026-02-15 20:38:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:38:28.585615759 +0000 UTC m=+1316.451631392" watchObservedRunningTime="2026-02-15 20:38:28.594476965 +0000 UTC m=+1316.460492588" Feb 15 20:38:28 crc kubenswrapper[4735]: I0215 20:38:28.625790 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.625771709 podStartE2EDuration="2.625771709s" podCreationTimestamp="2026-02-15 20:38:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:38:28.614752215 +0000 UTC m=+1316.480767848" watchObservedRunningTime="2026-02-15 20:38:28.625771709 +0000 UTC m=+1316.491787342" Feb 15 20:38:30 crc kubenswrapper[4735]: I0215 20:38:30.969270 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 15 20:38:30 crc kubenswrapper[4735]: I0215 20:38:30.969875 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 15 20:38:32 crc kubenswrapper[4735]: I0215 20:38:32.161826 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 15 20:38:34 crc kubenswrapper[4735]: I0215 20:38:34.778518 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5d466d479d-wdxj6" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 15 20:38:34 crc kubenswrapper[4735]: I0215 20:38:34.779075 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:38:35 crc kubenswrapper[4735]: I0215 20:38:35.969681 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 15 20:38:35 crc kubenswrapper[4735]: I0215 20:38:35.970220 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 15 20:38:36 crc kubenswrapper[4735]: I0215 20:38:36.982089 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="90c9c199-a27c-464a-8845-c2fd4e0b0fd9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 15 20:38:36 crc kubenswrapper[4735]: I0215 20:38:36.982177 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="90c9c199-a27c-464a-8845-c2fd4e0b0fd9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 15 20:38:37 crc kubenswrapper[4735]: I0215 20:38:37.132891 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 15 20:38:37 crc kubenswrapper[4735]: I0215 20:38:37.132990 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 15 20:38:37 crc kubenswrapper[4735]: I0215 20:38:37.162045 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 15 20:38:37 crc kubenswrapper[4735]: I0215 20:38:37.212525 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 15 20:38:37 crc kubenswrapper[4735]: I0215 20:38:37.746614 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 15 20:38:38 crc kubenswrapper[4735]: I0215 20:38:38.181170 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a964a23e-2012-4d43-98c2-e0b15f9d9d74" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 15 20:38:38 crc kubenswrapper[4735]: I0215 20:38:38.181751 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a964a23e-2012-4d43-98c2-e0b15f9d9d74" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.757446 4735 generic.go:334] "Generic (PLEG): container finished" podID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerID="5d80cb6d2dc2ccf6ac05974c0b49e10cfa35444bc235b804bde522b9f6bb18fb" exitCode=137 Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.758828 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerDied","Data":"5d80cb6d2dc2ccf6ac05974c0b49e10cfa35444bc235b804bde522b9f6bb18fb"} Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.758920 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d466d479d-wdxj6" event={"ID":"9f788fda-f0ae-45f7-a233-8e4963e5ed2d","Type":"ContainerDied","Data":"b1fff493b8f3f991759d71a9f7417a87ca0a4d350b5f0355919487363eeb3c7e"} Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.759000 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1fff493b8f3f991759d71a9f7417a87ca0a4d350b5f0355919487363eeb3c7e" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.764825 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.838066 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-logs\") pod \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.838196 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdsm4\" (UniqueName: \"kubernetes.io/projected/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-kube-api-access-xdsm4\") pod \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.838232 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-secret-key\") pod \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.838275 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-config-data\") pod \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.838379 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-tls-certs\") pod \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.838472 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-scripts\") pod \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.838513 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-combined-ca-bundle\") pod \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\" (UID: \"9f788fda-f0ae-45f7-a233-8e4963e5ed2d\") " Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.838586 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-logs" (OuterVolumeSpecName: "logs") pod "9f788fda-f0ae-45f7-a233-8e4963e5ed2d" (UID: "9f788fda-f0ae-45f7-a233-8e4963e5ed2d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.839038 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-logs\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.843604 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-kube-api-access-xdsm4" (OuterVolumeSpecName: "kube-api-access-xdsm4") pod "9f788fda-f0ae-45f7-a233-8e4963e5ed2d" (UID: "9f788fda-f0ae-45f7-a233-8e4963e5ed2d"). InnerVolumeSpecName "kube-api-access-xdsm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.852129 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9f788fda-f0ae-45f7-a233-8e4963e5ed2d" (UID: "9f788fda-f0ae-45f7-a233-8e4963e5ed2d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.868266 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-scripts" (OuterVolumeSpecName: "scripts") pod "9f788fda-f0ae-45f7-a233-8e4963e5ed2d" (UID: "9f788fda-f0ae-45f7-a233-8e4963e5ed2d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.868399 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f788fda-f0ae-45f7-a233-8e4963e5ed2d" (UID: "9f788fda-f0ae-45f7-a233-8e4963e5ed2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.868857 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-config-data" (OuterVolumeSpecName: "config-data") pod "9f788fda-f0ae-45f7-a233-8e4963e5ed2d" (UID: "9f788fda-f0ae-45f7-a233-8e4963e5ed2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.888149 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "9f788fda-f0ae-45f7-a233-8e4963e5ed2d" (UID: "9f788fda-f0ae-45f7-a233-8e4963e5ed2d"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.940630 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdsm4\" (UniqueName: \"kubernetes.io/projected/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-kube-api-access-xdsm4\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.940668 4735 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.940679 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.940687 4735 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.940696 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-scripts\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:41 crc kubenswrapper[4735]: I0215 20:38:41.940704 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f788fda-f0ae-45f7-a233-8e4963e5ed2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:38:42 crc kubenswrapper[4735]: I0215 20:38:42.770706 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d466d479d-wdxj6" Feb 15 20:38:42 crc kubenswrapper[4735]: I0215 20:38:42.809932 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d466d479d-wdxj6"] Feb 15 20:38:42 crc kubenswrapper[4735]: I0215 20:38:42.819903 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5d466d479d-wdxj6"] Feb 15 20:38:42 crc kubenswrapper[4735]: I0215 20:38:42.904160 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" path="/var/lib/kubelet/pods/9f788fda-f0ae-45f7-a233-8e4963e5ed2d/volumes" Feb 15 20:38:45 crc kubenswrapper[4735]: I0215 20:38:45.851313 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 15 20:38:45 crc kubenswrapper[4735]: I0215 20:38:45.975506 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 15 20:38:45 crc kubenswrapper[4735]: I0215 20:38:45.978924 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 15 20:38:45 crc kubenswrapper[4735]: I0215 20:38:45.986691 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 15 20:38:46 crc kubenswrapper[4735]: I0215 20:38:46.818280 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 15 20:38:47 crc kubenswrapper[4735]: I0215 20:38:47.140499 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 15 20:38:47 crc kubenswrapper[4735]: I0215 20:38:47.141398 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 15 20:38:47 crc kubenswrapper[4735]: I0215 20:38:47.148842 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 15 20:38:47 crc kubenswrapper[4735]: I0215 20:38:47.153551 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 15 20:38:47 crc kubenswrapper[4735]: I0215 20:38:47.821672 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 15 20:38:47 crc kubenswrapper[4735]: I0215 20:38:47.827754 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 15 20:38:49 crc kubenswrapper[4735]: I0215 20:38:49.680289 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:38:49 crc kubenswrapper[4735]: I0215 20:38:49.680666 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:38:55 crc kubenswrapper[4735]: I0215 20:38:55.550809 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 15 20:38:56 crc kubenswrapper[4735]: I0215 20:38:56.986953 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 15 20:39:05 crc kubenswrapper[4735]: I0215 20:39:05.134191 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="dc3c285a-910a-4cb1-968e-636216557d09" containerName="rabbitmq" containerID="cri-o://ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1" gracePeriod=604791 Feb 15 20:39:05 crc kubenswrapper[4735]: I0215 20:39:05.834140 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="5c7b91a3-10d6-465d-aff4-4855d697b54f" containerName="rabbitmq" containerID="cri-o://ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba" gracePeriod=604792 Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.457352 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619522 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-server-conf\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619563 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9727r\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-kube-api-access-9727r\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619588 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-plugins\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619609 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619713 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c7b91a3-10d6-465d-aff4-4855d697b54f-pod-info\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619729 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-confd\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619813 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-config-data\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619830 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-plugins-conf\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619847 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-erlang-cookie\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619863 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-tls\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.619900 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c7b91a3-10d6-465d-aff4-4855d697b54f-erlang-cookie-secret\") pod \"5c7b91a3-10d6-465d-aff4-4855d697b54f\" (UID: \"5c7b91a3-10d6-465d-aff4-4855d697b54f\") " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.621774 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.622478 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.624403 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.627276 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.627361 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c7b91a3-10d6-465d-aff4-4855d697b54f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.645234 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-kube-api-access-9727r" (OuterVolumeSpecName: "kube-api-access-9727r") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "kube-api-access-9727r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.645320 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.646743 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5c7b91a3-10d6-465d-aff4-4855d697b54f-pod-info" (OuterVolumeSpecName: "pod-info") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.653851 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-config-data" (OuterVolumeSpecName: "config-data") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.702506 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-server-conf" (OuterVolumeSpecName: "server-conf") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721688 4735 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c7b91a3-10d6-465d-aff4-4855d697b54f-pod-info\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721724 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721734 4735 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721744 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721754 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721763 4735 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c7b91a3-10d6-465d-aff4-4855d697b54f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721771 4735 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c7b91a3-10d6-465d-aff4-4855d697b54f-server-conf\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721779 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9727r\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-kube-api-access-9727r\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721786 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.721815 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.740528 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.761364 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5c7b91a3-10d6-465d-aff4-4855d697b54f" (UID: "5c7b91a3-10d6-465d-aff4-4855d697b54f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.823452 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:07 crc kubenswrapper[4735]: I0215 20:39:07.823495 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c7b91a3-10d6-465d-aff4-4855d697b54f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.053187 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c7b91a3-10d6-465d-aff4-4855d697b54f" containerID="ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba" exitCode=0 Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.053242 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5c7b91a3-10d6-465d-aff4-4855d697b54f","Type":"ContainerDied","Data":"ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba"} Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.053287 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.053311 4735 scope.go:117] "RemoveContainer" containerID="ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.053296 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5c7b91a3-10d6-465d-aff4-4855d697b54f","Type":"ContainerDied","Data":"6d058253789e53d5e75363ee2b4962a083c87a62523c4e2669f5d420caf190a5"} Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.086149 4735 scope.go:117] "RemoveContainer" containerID="528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.106035 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.118422 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.140075 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 15 20:39:08 crc kubenswrapper[4735]: E0215 20:39:08.141150 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon-log" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141170 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon-log" Feb 15 20:39:08 crc kubenswrapper[4735]: E0215 20:39:08.141187 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141197 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" Feb 15 20:39:08 crc kubenswrapper[4735]: E0215 20:39:08.141211 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c7b91a3-10d6-465d-aff4-4855d697b54f" containerName="setup-container" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141220 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c7b91a3-10d6-465d-aff4-4855d697b54f" containerName="setup-container" Feb 15 20:39:08 crc kubenswrapper[4735]: E0215 20:39:08.141242 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141250 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" Feb 15 20:39:08 crc kubenswrapper[4735]: E0215 20:39:08.141278 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141286 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" Feb 15 20:39:08 crc kubenswrapper[4735]: E0215 20:39:08.141312 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c7b91a3-10d6-465d-aff4-4855d697b54f" containerName="rabbitmq" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141320 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c7b91a3-10d6-465d-aff4-4855d697b54f" containerName="rabbitmq" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141509 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c7b91a3-10d6-465d-aff4-4855d697b54f" containerName="rabbitmq" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141521 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141534 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon-log" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141541 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.141551 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f788fda-f0ae-45f7-a233-8e4963e5ed2d" containerName="horizon" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.142512 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.146977 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.147039 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.147415 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-dvwjn" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.147667 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.147835 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.147891 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.157357 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.157876 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.165099 4735 scope.go:117] "RemoveContainer" containerID="ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba" Feb 15 20:39:08 crc kubenswrapper[4735]: E0215 20:39:08.165503 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba\": container with ID starting with ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba not found: ID does not exist" containerID="ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.165536 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba"} err="failed to get container status \"ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba\": rpc error: code = NotFound desc = could not find container \"ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba\": container with ID starting with ad8633173f8e75b8bff5c0ca8b65bf880417ae8a01e4ac232e27d2b20dd0fdba not found: ID does not exist" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.165557 4735 scope.go:117] "RemoveContainer" containerID="528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca" Feb 15 20:39:08 crc kubenswrapper[4735]: E0215 20:39:08.167929 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca\": container with ID starting with 528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca not found: ID does not exist" containerID="528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.167984 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca"} err="failed to get container status \"528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca\": rpc error: code = NotFound desc = could not find container \"528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca\": container with ID starting with 528697cc0ba05a014832df4f35ff3ee3e6a49377768751c6aeb6129358f93dca not found: ID does not exist" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.330489 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-c795m"] Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.332032 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333032 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333089 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1105ffb-a8a0-44fd-8679-171e016f43b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333106 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b1105ffb-a8a0-44fd-8679-171e016f43b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333128 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b1105ffb-a8a0-44fd-8679-171e016f43b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333174 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b1105ffb-a8a0-44fd-8679-171e016f43b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333205 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333238 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333323 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333350 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b1105ffb-a8a0-44fd-8679-171e016f43b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333371 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wz2q\" (UniqueName: \"kubernetes.io/projected/b1105ffb-a8a0-44fd-8679-171e016f43b1-kube-api-access-4wz2q\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.333396 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.335578 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.344223 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-c795m"] Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.434768 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.434818 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-config\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.434853 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz8kd\" (UniqueName: \"kubernetes.io/projected/cddd81fe-9191-4a6c-8eba-4956dc88f06d-kube-api-access-dz8kd\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.434877 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.434899 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.434920 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-svc\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.434939 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.434995 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1105ffb-a8a0-44fd-8679-171e016f43b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435013 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b1105ffb-a8a0-44fd-8679-171e016f43b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435031 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b1105ffb-a8a0-44fd-8679-171e016f43b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435065 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435093 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b1105ffb-a8a0-44fd-8679-171e016f43b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435115 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435138 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435172 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435196 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435216 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b1105ffb-a8a0-44fd-8679-171e016f43b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435235 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wz2q\" (UniqueName: \"kubernetes.io/projected/b1105ffb-a8a0-44fd-8679-171e016f43b1-kube-api-access-4wz2q\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.435319 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.436027 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.436306 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.437101 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1105ffb-a8a0-44fd-8679-171e016f43b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.437236 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b1105ffb-a8a0-44fd-8679-171e016f43b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.437742 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b1105ffb-a8a0-44fd-8679-171e016f43b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.440522 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.442421 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b1105ffb-a8a0-44fd-8679-171e016f43b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.442757 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b1105ffb-a8a0-44fd-8679-171e016f43b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.443557 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b1105ffb-a8a0-44fd-8679-171e016f43b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.452090 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wz2q\" (UniqueName: \"kubernetes.io/projected/b1105ffb-a8a0-44fd-8679-171e016f43b1-kube-api-access-4wz2q\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.479552 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1105ffb-a8a0-44fd-8679-171e016f43b1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.513840 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.537444 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-config\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.537497 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz8kd\" (UniqueName: \"kubernetes.io/projected/cddd81fe-9191-4a6c-8eba-4956dc88f06d-kube-api-access-dz8kd\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.537522 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.537543 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.537561 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-svc\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.537627 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.537658 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.538375 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-svc\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.538438 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-config\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.540468 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.540705 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.540744 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.541224 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.558210 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz8kd\" (UniqueName: \"kubernetes.io/projected/cddd81fe-9191-4a6c-8eba-4956dc88f06d-kube-api-access-dz8kd\") pod \"dnsmasq-dns-d558885bc-c795m\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.647672 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:08 crc kubenswrapper[4735]: I0215 20:39:08.900432 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c7b91a3-10d6-465d-aff4-4855d697b54f" path="/var/lib/kubelet/pods/5c7b91a3-10d6-465d-aff4-4855d697b54f/volumes" Feb 15 20:39:09 crc kubenswrapper[4735]: I0215 20:39:09.045181 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 15 20:39:09 crc kubenswrapper[4735]: I0215 20:39:09.063865 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b1105ffb-a8a0-44fd-8679-171e016f43b1","Type":"ContainerStarted","Data":"962a1cadc1b5d4543d5ff6e64ba93b0c17e059d728ddc0fe1f40a06b80ec1a12"} Feb 15 20:39:09 crc kubenswrapper[4735]: I0215 20:39:09.164219 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-c795m"] Feb 15 20:39:09 crc kubenswrapper[4735]: W0215 20:39:09.168976 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcddd81fe_9191_4a6c_8eba_4956dc88f06d.slice/crio-5e2b52f146d9d5c38d2b5a5ac79f931208e7e1b52a5c4128298b15f39b9dbcb7 WatchSource:0}: Error finding container 5e2b52f146d9d5c38d2b5a5ac79f931208e7e1b52a5c4128298b15f39b9dbcb7: Status 404 returned error can't find the container with id 5e2b52f146d9d5c38d2b5a5ac79f931208e7e1b52a5c4128298b15f39b9dbcb7 Feb 15 20:39:10 crc kubenswrapper[4735]: I0215 20:39:10.075798 4735 generic.go:334] "Generic (PLEG): container finished" podID="cddd81fe-9191-4a6c-8eba-4956dc88f06d" containerID="65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e" exitCode=0 Feb 15 20:39:10 crc kubenswrapper[4735]: I0215 20:39:10.075878 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-c795m" event={"ID":"cddd81fe-9191-4a6c-8eba-4956dc88f06d","Type":"ContainerDied","Data":"65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e"} Feb 15 20:39:10 crc kubenswrapper[4735]: I0215 20:39:10.077303 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-c795m" event={"ID":"cddd81fe-9191-4a6c-8eba-4956dc88f06d","Type":"ContainerStarted","Data":"5e2b52f146d9d5c38d2b5a5ac79f931208e7e1b52a5c4128298b15f39b9dbcb7"} Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.072381 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="dc3c285a-910a-4cb1-968e-636216557d09" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.091529 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b1105ffb-a8a0-44fd-8679-171e016f43b1","Type":"ContainerStarted","Data":"bb81f4627af936e924abd7402a62c49e22d06a024ea8f29879f6e0abad8f7671"} Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.097038 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-c795m" event={"ID":"cddd81fe-9191-4a6c-8eba-4956dc88f06d","Type":"ContainerStarted","Data":"1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9"} Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.097532 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.163179 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-c795m" podStartSLOduration=3.163155651 podStartE2EDuration="3.163155651s" podCreationTimestamp="2026-02-15 20:39:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:39:11.150360389 +0000 UTC m=+1359.016376032" watchObservedRunningTime="2026-02-15 20:39:11.163155651 +0000 UTC m=+1359.029171284" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.697202 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807397 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-plugins\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807462 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-confd\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807509 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc3c285a-910a-4cb1-968e-636216557d09-pod-info\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807547 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-plugins-conf\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807585 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52cqz\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-kube-api-access-52cqz\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807679 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc3c285a-910a-4cb1-968e-636216557d09-erlang-cookie-secret\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807754 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807782 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-tls\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807801 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-config-data\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807841 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-erlang-cookie\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.807865 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-server-conf\") pod \"dc3c285a-910a-4cb1-968e-636216557d09\" (UID: \"dc3c285a-910a-4cb1-968e-636216557d09\") " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.808370 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.808669 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.811401 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.814580 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.817141 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.817421 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/dc3c285a-910a-4cb1-968e-636216557d09-pod-info" (OuterVolumeSpecName: "pod-info") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.821436 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.825621 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc3c285a-910a-4cb1-968e-636216557d09-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.835331 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-kube-api-access-52cqz" (OuterVolumeSpecName: "kube-api-access-52cqz") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "kube-api-access-52cqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.856452 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-config-data" (OuterVolumeSpecName: "config-data") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.910473 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52cqz\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-kube-api-access-52cqz\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.910501 4735 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc3c285a-910a-4cb1-968e-636216557d09-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.910530 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.910540 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.910550 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.910557 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.910565 4735 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc3c285a-910a-4cb1-968e-636216557d09-pod-info\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.910577 4735 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.911185 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-server-conf" (OuterVolumeSpecName: "server-conf") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.936898 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "dc3c285a-910a-4cb1-968e-636216557d09" (UID: "dc3c285a-910a-4cb1-968e-636216557d09"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:39:11 crc kubenswrapper[4735]: I0215 20:39:11.939977 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.012058 4735 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc3c285a-910a-4cb1-968e-636216557d09-server-conf\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.012324 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc3c285a-910a-4cb1-968e-636216557d09-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.012421 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.108264 4735 generic.go:334] "Generic (PLEG): container finished" podID="dc3c285a-910a-4cb1-968e-636216557d09" containerID="ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1" exitCode=0 Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.108365 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.108997 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc3c285a-910a-4cb1-968e-636216557d09","Type":"ContainerDied","Data":"ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1"} Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.109036 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc3c285a-910a-4cb1-968e-636216557d09","Type":"ContainerDied","Data":"6aed3fa6dea6824209d54628ebf241ba43a0520b0dfd95cfc059dd181621c81c"} Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.109054 4735 scope.go:117] "RemoveContainer" containerID="ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.137372 4735 scope.go:117] "RemoveContainer" containerID="6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.141005 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.156628 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.175110 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 15 20:39:12 crc kubenswrapper[4735]: E0215 20:39:12.175505 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3c285a-910a-4cb1-968e-636216557d09" containerName="setup-container" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.175517 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3c285a-910a-4cb1-968e-636216557d09" containerName="setup-container" Feb 15 20:39:12 crc kubenswrapper[4735]: E0215 20:39:12.175530 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3c285a-910a-4cb1-968e-636216557d09" containerName="rabbitmq" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.175536 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3c285a-910a-4cb1-968e-636216557d09" containerName="rabbitmq" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.175688 4735 scope.go:117] "RemoveContainer" containerID="ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.175736 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc3c285a-910a-4cb1-968e-636216557d09" containerName="rabbitmq" Feb 15 20:39:12 crc kubenswrapper[4735]: E0215 20:39:12.176771 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1\": container with ID starting with ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1 not found: ID does not exist" containerID="ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.176804 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1"} err="failed to get container status \"ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1\": rpc error: code = NotFound desc = could not find container \"ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1\": container with ID starting with ae5b9fdbe4ab32d9b17b63554be0cb3d397758422d91d9cd3d6fb81567a7a1c1 not found: ID does not exist" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.176837 4735 scope.go:117] "RemoveContainer" containerID="6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.176818 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: E0215 20:39:12.178803 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a\": container with ID starting with 6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a not found: ID does not exist" containerID="6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.178841 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a"} err="failed to get container status \"6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a\": rpc error: code = NotFound desc = could not find container \"6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a\": container with ID starting with 6927741086d17471682aedddf5f2d3085c01384ceb69c3170d7d3754547c9c0a not found: ID does not exist" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.181013 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.181357 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.181401 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-g2d24" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.181426 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.181635 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.181655 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.192745 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.227960 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323216 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323261 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323508 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-config-data\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323671 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323777 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323807 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg7fs\" (UniqueName: \"kubernetes.io/projected/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-kube-api-access-kg7fs\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323850 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323890 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323919 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323965 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.323991 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425299 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425361 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425397 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425417 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425435 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425455 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425502 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-config-data\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425548 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425581 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425597 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg7fs\" (UniqueName: \"kubernetes.io/projected/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-kube-api-access-kg7fs\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425618 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.425894 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.426155 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.426184 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.427457 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-config-data\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.427732 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.429647 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.430965 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.431422 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.431842 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.450436 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.466778 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg7fs\" (UniqueName: \"kubernetes.io/projected/ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40-kube-api-access-kg7fs\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.523351 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40\") " pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.810312 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 15 20:39:12 crc kubenswrapper[4735]: I0215 20:39:12.910209 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc3c285a-910a-4cb1-968e-636216557d09" path="/var/lib/kubelet/pods/dc3c285a-910a-4cb1-968e-636216557d09/volumes" Feb 15 20:39:13 crc kubenswrapper[4735]: I0215 20:39:13.284155 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 15 20:39:14 crc kubenswrapper[4735]: I0215 20:39:14.144001 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40","Type":"ContainerStarted","Data":"09f8033e737761951bd84bfbef206e648df2b9639ecbab23136a7bf81ec89790"} Feb 15 20:39:15 crc kubenswrapper[4735]: I0215 20:39:15.153768 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40","Type":"ContainerStarted","Data":"5db05ce311c073379348e42d1c8a649582d631594eaa4c52dcd11bbbdc01833c"} Feb 15 20:39:18 crc kubenswrapper[4735]: I0215 20:39:18.650356 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:18 crc kubenswrapper[4735]: I0215 20:39:18.712489 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-892xq"] Feb 15 20:39:18 crc kubenswrapper[4735]: I0215 20:39:18.712720 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" podUID="100107cc-9e89-476e-a44c-ca88b04793cc" containerName="dnsmasq-dns" containerID="cri-o://559c831cd9849fbdc2e511ca9fbec70fdddb93b5598fef560746a9ec1621c018" gracePeriod=10 Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.095638 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-798f84bf5f-pnv2f"] Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.097368 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.149201 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798f84bf5f-pnv2f"] Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.170988 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-dns-svc\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.171036 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-config\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.171060 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbb8h\" (UniqueName: \"kubernetes.io/projected/fd43082b-0db8-4620-9784-cc587548f737-kube-api-access-xbb8h\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.171136 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-dns-swift-storage-0\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.171152 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-ovsdbserver-sb\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.171181 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-openstack-edpm-ipam\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.171202 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-ovsdbserver-nb\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.211266 4735 generic.go:334] "Generic (PLEG): container finished" podID="100107cc-9e89-476e-a44c-ca88b04793cc" containerID="559c831cd9849fbdc2e511ca9fbec70fdddb93b5598fef560746a9ec1621c018" exitCode=0 Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.211572 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" event={"ID":"100107cc-9e89-476e-a44c-ca88b04793cc","Type":"ContainerDied","Data":"559c831cd9849fbdc2e511ca9fbec70fdddb93b5598fef560746a9ec1621c018"} Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.273129 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-config\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.273171 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbb8h\" (UniqueName: \"kubernetes.io/projected/fd43082b-0db8-4620-9784-cc587548f737-kube-api-access-xbb8h\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.273257 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-ovsdbserver-sb\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.273272 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-dns-swift-storage-0\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.273306 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-openstack-edpm-ipam\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.273328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-ovsdbserver-nb\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.273389 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-dns-svc\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.277598 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-dns-svc\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.279417 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-ovsdbserver-sb\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.291683 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-dns-swift-storage-0\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.297487 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-config\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.298569 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-ovsdbserver-nb\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.301025 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fd43082b-0db8-4620-9784-cc587548f737-openstack-edpm-ipam\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.329786 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbb8h\" (UniqueName: \"kubernetes.io/projected/fd43082b-0db8-4620-9784-cc587548f737-kube-api-access-xbb8h\") pod \"dnsmasq-dns-798f84bf5f-pnv2f\" (UID: \"fd43082b-0db8-4620-9784-cc587548f737\") " pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.409874 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.422709 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.480512 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-config\") pod \"100107cc-9e89-476e-a44c-ca88b04793cc\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.480644 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ws2k\" (UniqueName: \"kubernetes.io/projected/100107cc-9e89-476e-a44c-ca88b04793cc-kube-api-access-7ws2k\") pod \"100107cc-9e89-476e-a44c-ca88b04793cc\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.480760 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-nb\") pod \"100107cc-9e89-476e-a44c-ca88b04793cc\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.480806 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-swift-storage-0\") pod \"100107cc-9e89-476e-a44c-ca88b04793cc\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.480829 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-sb\") pod \"100107cc-9e89-476e-a44c-ca88b04793cc\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.480934 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-svc\") pod \"100107cc-9e89-476e-a44c-ca88b04793cc\" (UID: \"100107cc-9e89-476e-a44c-ca88b04793cc\") " Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.499105 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/100107cc-9e89-476e-a44c-ca88b04793cc-kube-api-access-7ws2k" (OuterVolumeSpecName: "kube-api-access-7ws2k") pod "100107cc-9e89-476e-a44c-ca88b04793cc" (UID: "100107cc-9e89-476e-a44c-ca88b04793cc"). InnerVolumeSpecName "kube-api-access-7ws2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.546235 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-config" (OuterVolumeSpecName: "config") pod "100107cc-9e89-476e-a44c-ca88b04793cc" (UID: "100107cc-9e89-476e-a44c-ca88b04793cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.572376 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "100107cc-9e89-476e-a44c-ca88b04793cc" (UID: "100107cc-9e89-476e-a44c-ca88b04793cc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.584559 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "100107cc-9e89-476e-a44c-ca88b04793cc" (UID: "100107cc-9e89-476e-a44c-ca88b04793cc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.584778 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.584800 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.584810 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.584818 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ws2k\" (UniqueName: \"kubernetes.io/projected/100107cc-9e89-476e-a44c-ca88b04793cc-kube-api-access-7ws2k\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.591838 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "100107cc-9e89-476e-a44c-ca88b04793cc" (UID: "100107cc-9e89-476e-a44c-ca88b04793cc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.625760 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "100107cc-9e89-476e-a44c-ca88b04793cc" (UID: "100107cc-9e89-476e-a44c-ca88b04793cc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.681193 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.681254 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.690462 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.690508 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/100107cc-9e89-476e-a44c-ca88b04793cc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:19 crc kubenswrapper[4735]: I0215 20:39:19.918014 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798f84bf5f-pnv2f"] Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.221308 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" event={"ID":"100107cc-9e89-476e-a44c-ca88b04793cc","Type":"ContainerDied","Data":"79e68b0b3d0355cea189ef133dc7745eb1004520bd8573b28cecd509118aba6b"} Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.222507 4735 scope.go:117] "RemoveContainer" containerID="559c831cd9849fbdc2e511ca9fbec70fdddb93b5598fef560746a9ec1621c018" Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.221355 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-892xq" Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.223006 4735 generic.go:334] "Generic (PLEG): container finished" podID="fd43082b-0db8-4620-9784-cc587548f737" containerID="7c30c78c18a641c765d326083dd70c76afe8a81da417250248dddfb60d0302c0" exitCode=0 Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.223050 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" event={"ID":"fd43082b-0db8-4620-9784-cc587548f737","Type":"ContainerDied","Data":"7c30c78c18a641c765d326083dd70c76afe8a81da417250248dddfb60d0302c0"} Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.223073 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" event={"ID":"fd43082b-0db8-4620-9784-cc587548f737","Type":"ContainerStarted","Data":"b1901db0f737e61ed343e6616a12c74ddc4b6bbd303d6e174a0dd2b899a7b244"} Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.273419 4735 scope.go:117] "RemoveContainer" containerID="2c24fedc4a4f755da0d3ea38c9cbfae50208025fdb6719a9009bcd932a613aa3" Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.289277 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-892xq"] Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.297305 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-892xq"] Feb 15 20:39:20 crc kubenswrapper[4735]: I0215 20:39:20.899616 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="100107cc-9e89-476e-a44c-ca88b04793cc" path="/var/lib/kubelet/pods/100107cc-9e89-476e-a44c-ca88b04793cc/volumes" Feb 15 20:39:21 crc kubenswrapper[4735]: I0215 20:39:21.232765 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" event={"ID":"fd43082b-0db8-4620-9784-cc587548f737","Type":"ContainerStarted","Data":"6052fe7c75ed0b9043296042982408615df30aa696de573cc87b3d7d8830577d"} Feb 15 20:39:21 crc kubenswrapper[4735]: I0215 20:39:21.232874 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:21 crc kubenswrapper[4735]: I0215 20:39:21.258996 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" podStartSLOduration=2.258978801 podStartE2EDuration="2.258978801s" podCreationTimestamp="2026-02-15 20:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:39:21.249805887 +0000 UTC m=+1369.115821520" watchObservedRunningTime="2026-02-15 20:39:21.258978801 +0000 UTC m=+1369.124994424" Feb 15 20:39:29 crc kubenswrapper[4735]: I0215 20:39:29.425182 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-798f84bf5f-pnv2f" Feb 15 20:39:29 crc kubenswrapper[4735]: I0215 20:39:29.513900 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-c795m"] Feb 15 20:39:29 crc kubenswrapper[4735]: I0215 20:39:29.514185 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-c795m" podUID="cddd81fe-9191-4a6c-8eba-4956dc88f06d" containerName="dnsmasq-dns" containerID="cri-o://1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9" gracePeriod=10 Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.032532 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.201165 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-config\") pod \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.201477 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz8kd\" (UniqueName: \"kubernetes.io/projected/cddd81fe-9191-4a6c-8eba-4956dc88f06d-kube-api-access-dz8kd\") pod \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.201533 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-nb\") pod \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.201557 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-sb\") pod \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.201733 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-openstack-edpm-ipam\") pod \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.201782 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-svc\") pod \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.201813 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-swift-storage-0\") pod \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\" (UID: \"cddd81fe-9191-4a6c-8eba-4956dc88f06d\") " Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.217148 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cddd81fe-9191-4a6c-8eba-4956dc88f06d-kube-api-access-dz8kd" (OuterVolumeSpecName: "kube-api-access-dz8kd") pod "cddd81fe-9191-4a6c-8eba-4956dc88f06d" (UID: "cddd81fe-9191-4a6c-8eba-4956dc88f06d"). InnerVolumeSpecName "kube-api-access-dz8kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.246358 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "cddd81fe-9191-4a6c-8eba-4956dc88f06d" (UID: "cddd81fe-9191-4a6c-8eba-4956dc88f06d"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.250747 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-config" (OuterVolumeSpecName: "config") pod "cddd81fe-9191-4a6c-8eba-4956dc88f06d" (UID: "cddd81fe-9191-4a6c-8eba-4956dc88f06d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.256180 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cddd81fe-9191-4a6c-8eba-4956dc88f06d" (UID: "cddd81fe-9191-4a6c-8eba-4956dc88f06d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.260319 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cddd81fe-9191-4a6c-8eba-4956dc88f06d" (UID: "cddd81fe-9191-4a6c-8eba-4956dc88f06d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.269937 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cddd81fe-9191-4a6c-8eba-4956dc88f06d" (UID: "cddd81fe-9191-4a6c-8eba-4956dc88f06d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.272505 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cddd81fe-9191-4a6c-8eba-4956dc88f06d" (UID: "cddd81fe-9191-4a6c-8eba-4956dc88f06d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.307190 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.307219 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.307229 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.307241 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.307252 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.307262 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cddd81fe-9191-4a6c-8eba-4956dc88f06d-config\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.307271 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz8kd\" (UniqueName: \"kubernetes.io/projected/cddd81fe-9191-4a6c-8eba-4956dc88f06d-kube-api-access-dz8kd\") on node \"crc\" DevicePath \"\"" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.322719 4735 generic.go:334] "Generic (PLEG): container finished" podID="cddd81fe-9191-4a6c-8eba-4956dc88f06d" containerID="1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9" exitCode=0 Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.322754 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-c795m" event={"ID":"cddd81fe-9191-4a6c-8eba-4956dc88f06d","Type":"ContainerDied","Data":"1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9"} Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.322779 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-c795m" event={"ID":"cddd81fe-9191-4a6c-8eba-4956dc88f06d","Type":"ContainerDied","Data":"5e2b52f146d9d5c38d2b5a5ac79f931208e7e1b52a5c4128298b15f39b9dbcb7"} Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.322795 4735 scope.go:117] "RemoveContainer" containerID="1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.322901 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-c795m" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.343202 4735 scope.go:117] "RemoveContainer" containerID="65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.354860 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-c795m"] Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.362486 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-c795m"] Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.370321 4735 scope.go:117] "RemoveContainer" containerID="1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9" Feb 15 20:39:30 crc kubenswrapper[4735]: E0215 20:39:30.377168 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9\": container with ID starting with 1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9 not found: ID does not exist" containerID="1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.377215 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9"} err="failed to get container status \"1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9\": rpc error: code = NotFound desc = could not find container \"1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9\": container with ID starting with 1223c9f27a38693e7fdaaaeb93d9184c28518eac10197d37dd11bdafd5114dc9 not found: ID does not exist" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.377241 4735 scope.go:117] "RemoveContainer" containerID="65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e" Feb 15 20:39:30 crc kubenswrapper[4735]: E0215 20:39:30.377611 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e\": container with ID starting with 65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e not found: ID does not exist" containerID="65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.377629 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e"} err="failed to get container status \"65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e\": rpc error: code = NotFound desc = could not find container \"65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e\": container with ID starting with 65f4333684babc4824fcd3f729244b8dd2fce916c763d8f4c73d0d9c68a5d81e not found: ID does not exist" Feb 15 20:39:30 crc kubenswrapper[4735]: I0215 20:39:30.899095 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cddd81fe-9191-4a6c-8eba-4956dc88f06d" path="/var/lib/kubelet/pods/cddd81fe-9191-4a6c-8eba-4956dc88f06d/volumes" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.445465 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nbxhf"] Feb 15 20:39:33 crc kubenswrapper[4735]: E0215 20:39:33.447184 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cddd81fe-9191-4a6c-8eba-4956dc88f06d" containerName="dnsmasq-dns" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.447303 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="cddd81fe-9191-4a6c-8eba-4956dc88f06d" containerName="dnsmasq-dns" Feb 15 20:39:33 crc kubenswrapper[4735]: E0215 20:39:33.447412 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="100107cc-9e89-476e-a44c-ca88b04793cc" containerName="dnsmasq-dns" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.447495 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="100107cc-9e89-476e-a44c-ca88b04793cc" containerName="dnsmasq-dns" Feb 15 20:39:33 crc kubenswrapper[4735]: E0215 20:39:33.447600 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cddd81fe-9191-4a6c-8eba-4956dc88f06d" containerName="init" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.447705 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="cddd81fe-9191-4a6c-8eba-4956dc88f06d" containerName="init" Feb 15 20:39:33 crc kubenswrapper[4735]: E0215 20:39:33.447795 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="100107cc-9e89-476e-a44c-ca88b04793cc" containerName="init" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.447879 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="100107cc-9e89-476e-a44c-ca88b04793cc" containerName="init" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.448223 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="100107cc-9e89-476e-a44c-ca88b04793cc" containerName="dnsmasq-dns" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.448356 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="cddd81fe-9191-4a6c-8eba-4956dc88f06d" containerName="dnsmasq-dns" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.450977 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.464054 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nbxhf"] Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.571375 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-catalog-content\") pod \"redhat-operators-nbxhf\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.571628 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-utilities\") pod \"redhat-operators-nbxhf\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.571658 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mwm9\" (UniqueName: \"kubernetes.io/projected/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-kube-api-access-9mwm9\") pod \"redhat-operators-nbxhf\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.674221 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-catalog-content\") pod \"redhat-operators-nbxhf\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.674485 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-utilities\") pod \"redhat-operators-nbxhf\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.674596 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mwm9\" (UniqueName: \"kubernetes.io/projected/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-kube-api-access-9mwm9\") pod \"redhat-operators-nbxhf\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.674727 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-catalog-content\") pod \"redhat-operators-nbxhf\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.675120 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-utilities\") pod \"redhat-operators-nbxhf\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.708994 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mwm9\" (UniqueName: \"kubernetes.io/projected/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-kube-api-access-9mwm9\") pod \"redhat-operators-nbxhf\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:33 crc kubenswrapper[4735]: I0215 20:39:33.778625 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:34 crc kubenswrapper[4735]: I0215 20:39:34.314005 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nbxhf"] Feb 15 20:39:34 crc kubenswrapper[4735]: I0215 20:39:34.359330 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbxhf" event={"ID":"fef0cbc1-3bb0-49e2-a5d0-4680f869c150","Type":"ContainerStarted","Data":"8a41def63a801d9d985ae6f99d64a623a0c9e0d566bb47f12e849ade5fbf4363"} Feb 15 20:39:35 crc kubenswrapper[4735]: I0215 20:39:35.368185 4735 generic.go:334] "Generic (PLEG): container finished" podID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerID="770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502" exitCode=0 Feb 15 20:39:35 crc kubenswrapper[4735]: I0215 20:39:35.368285 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbxhf" event={"ID":"fef0cbc1-3bb0-49e2-a5d0-4680f869c150","Type":"ContainerDied","Data":"770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502"} Feb 15 20:39:36 crc kubenswrapper[4735]: I0215 20:39:36.378347 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbxhf" event={"ID":"fef0cbc1-3bb0-49e2-a5d0-4680f869c150","Type":"ContainerStarted","Data":"376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b"} Feb 15 20:39:41 crc kubenswrapper[4735]: I0215 20:39:41.422511 4735 generic.go:334] "Generic (PLEG): container finished" podID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerID="376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b" exitCode=0 Feb 15 20:39:41 crc kubenswrapper[4735]: I0215 20:39:41.422553 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbxhf" event={"ID":"fef0cbc1-3bb0-49e2-a5d0-4680f869c150","Type":"ContainerDied","Data":"376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b"} Feb 15 20:39:42 crc kubenswrapper[4735]: I0215 20:39:42.432057 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbxhf" event={"ID":"fef0cbc1-3bb0-49e2-a5d0-4680f869c150","Type":"ContainerStarted","Data":"0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679"} Feb 15 20:39:42 crc kubenswrapper[4735]: I0215 20:39:42.466476 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nbxhf" podStartSLOduration=2.713245203 podStartE2EDuration="9.466456127s" podCreationTimestamp="2026-02-15 20:39:33 +0000 UTC" firstStartedPulling="2026-02-15 20:39:35.370328093 +0000 UTC m=+1383.236343716" lastFinishedPulling="2026-02-15 20:39:42.123538977 +0000 UTC m=+1389.989554640" observedRunningTime="2026-02-15 20:39:42.460399125 +0000 UTC m=+1390.326414748" watchObservedRunningTime="2026-02-15 20:39:42.466456127 +0000 UTC m=+1390.332471750" Feb 15 20:39:43 crc kubenswrapper[4735]: I0215 20:39:43.457918 4735 generic.go:334] "Generic (PLEG): container finished" podID="b1105ffb-a8a0-44fd-8679-171e016f43b1" containerID="bb81f4627af936e924abd7402a62c49e22d06a024ea8f29879f6e0abad8f7671" exitCode=0 Feb 15 20:39:43 crc kubenswrapper[4735]: I0215 20:39:43.458259 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b1105ffb-a8a0-44fd-8679-171e016f43b1","Type":"ContainerDied","Data":"bb81f4627af936e924abd7402a62c49e22d06a024ea8f29879f6e0abad8f7671"} Feb 15 20:39:43 crc kubenswrapper[4735]: I0215 20:39:43.779716 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:43 crc kubenswrapper[4735]: I0215 20:39:43.780153 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:39:44 crc kubenswrapper[4735]: I0215 20:39:44.468730 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b1105ffb-a8a0-44fd-8679-171e016f43b1","Type":"ContainerStarted","Data":"747bacf1e072c97016272df94aaba58bf867ca559d926d232cb35083d0de5b8e"} Feb 15 20:39:44 crc kubenswrapper[4735]: I0215 20:39:44.468967 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:39:44 crc kubenswrapper[4735]: I0215 20:39:44.509807 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.509788814 podStartE2EDuration="36.509788814s" podCreationTimestamp="2026-02-15 20:39:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:39:44.504983486 +0000 UTC m=+1392.370999119" watchObservedRunningTime="2026-02-15 20:39:44.509788814 +0000 UTC m=+1392.375804437" Feb 15 20:39:44 crc kubenswrapper[4735]: I0215 20:39:44.829973 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nbxhf" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="registry-server" probeResult="failure" output=< Feb 15 20:39:44 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:39:44 crc kubenswrapper[4735]: > Feb 15 20:39:47 crc kubenswrapper[4735]: I0215 20:39:47.493348 4735 generic.go:334] "Generic (PLEG): container finished" podID="ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40" containerID="5db05ce311c073379348e42d1c8a649582d631594eaa4c52dcd11bbbdc01833c" exitCode=0 Feb 15 20:39:47 crc kubenswrapper[4735]: I0215 20:39:47.493430 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40","Type":"ContainerDied","Data":"5db05ce311c073379348e42d1c8a649582d631594eaa4c52dcd11bbbdc01833c"} Feb 15 20:39:48 crc kubenswrapper[4735]: I0215 20:39:48.504053 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40","Type":"ContainerStarted","Data":"acd210fab0786b51e7c0bacf7994f38902c765389ec5f3545f03d5153a3a7d68"} Feb 15 20:39:48 crc kubenswrapper[4735]: I0215 20:39:48.504572 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 15 20:39:48 crc kubenswrapper[4735]: I0215 20:39:48.531598 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.531579332 podStartE2EDuration="36.531579332s" podCreationTimestamp="2026-02-15 20:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:39:48.524674958 +0000 UTC m=+1396.390690581" watchObservedRunningTime="2026-02-15 20:39:48.531579332 +0000 UTC m=+1396.397594955" Feb 15 20:39:49 crc kubenswrapper[4735]: I0215 20:39:49.679718 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:39:49 crc kubenswrapper[4735]: I0215 20:39:49.680095 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:39:49 crc kubenswrapper[4735]: I0215 20:39:49.680143 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:39:49 crc kubenswrapper[4735]: I0215 20:39:49.680861 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97239e5cb100e71e3f44686a5d1c9c6d1a3b14c88d00878cdd46ec7dbf229a42"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:39:49 crc kubenswrapper[4735]: I0215 20:39:49.680916 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://97239e5cb100e71e3f44686a5d1c9c6d1a3b14c88d00878cdd46ec7dbf229a42" gracePeriod=600 Feb 15 20:39:50 crc kubenswrapper[4735]: I0215 20:39:50.522540 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="97239e5cb100e71e3f44686a5d1c9c6d1a3b14c88d00878cdd46ec7dbf229a42" exitCode=0 Feb 15 20:39:50 crc kubenswrapper[4735]: I0215 20:39:50.522623 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"97239e5cb100e71e3f44686a5d1c9c6d1a3b14c88d00878cdd46ec7dbf229a42"} Feb 15 20:39:50 crc kubenswrapper[4735]: I0215 20:39:50.522816 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e"} Feb 15 20:39:50 crc kubenswrapper[4735]: I0215 20:39:50.522838 4735 scope.go:117] "RemoveContainer" containerID="94228183c3d1abcbd5b788cf2f92ec757e11f8d18a65dfdd44912dcfc7991e88" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.617351 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2"] Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.618753 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.621731 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.622046 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.622240 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.622368 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.662971 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.663036 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.663123 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stdhj\" (UniqueName: \"kubernetes.io/projected/e31b2a4c-8571-4a0e-868f-f86b757785ac-kube-api-access-stdhj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.663166 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.673293 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2"] Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.764870 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stdhj\" (UniqueName: \"kubernetes.io/projected/e31b2a4c-8571-4a0e-868f-f86b757785ac-kube-api-access-stdhj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.764940 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.765018 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.765060 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.781924 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.782097 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.784578 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.801766 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stdhj\" (UniqueName: \"kubernetes.io/projected/e31b2a4c-8571-4a0e-868f-f86b757785ac-kube-api-access-stdhj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:52 crc kubenswrapper[4735]: I0215 20:39:52.952707 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:39:53 crc kubenswrapper[4735]: I0215 20:39:53.790728 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2"] Feb 15 20:39:53 crc kubenswrapper[4735]: W0215 20:39:53.807536 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode31b2a4c_8571_4a0e_868f_f86b757785ac.slice/crio-beedb66ff5e1f276cd59355eafe3e5cfad2973819f2221161b101135fb190ea6 WatchSource:0}: Error finding container beedb66ff5e1f276cd59355eafe3e5cfad2973819f2221161b101135fb190ea6: Status 404 returned error can't find the container with id beedb66ff5e1f276cd59355eafe3e5cfad2973819f2221161b101135fb190ea6 Feb 15 20:39:54 crc kubenswrapper[4735]: I0215 20:39:54.560684 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" event={"ID":"e31b2a4c-8571-4a0e-868f-f86b757785ac","Type":"ContainerStarted","Data":"beedb66ff5e1f276cd59355eafe3e5cfad2973819f2221161b101135fb190ea6"} Feb 15 20:39:54 crc kubenswrapper[4735]: I0215 20:39:54.849991 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nbxhf" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="registry-server" probeResult="failure" output=< Feb 15 20:39:54 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:39:54 crc kubenswrapper[4735]: > Feb 15 20:39:58 crc kubenswrapper[4735]: I0215 20:39:58.518511 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 15 20:40:02 crc kubenswrapper[4735]: I0215 20:40:02.814087 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 15 20:40:04 crc kubenswrapper[4735]: I0215 20:40:04.853980 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nbxhf" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="registry-server" probeResult="failure" output=< Feb 15 20:40:04 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:40:04 crc kubenswrapper[4735]: > Feb 15 20:40:06 crc kubenswrapper[4735]: I0215 20:40:06.702209 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" event={"ID":"e31b2a4c-8571-4a0e-868f-f86b757785ac","Type":"ContainerStarted","Data":"71989ec6b9c1f2ab3e1a00951d7ef430245a06736e62846f7266b8dbd7b3e77a"} Feb 15 20:40:06 crc kubenswrapper[4735]: I0215 20:40:06.736071 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" podStartSLOduration=2.8774496000000003 podStartE2EDuration="14.736049571s" podCreationTimestamp="2026-02-15 20:39:52 +0000 UTC" firstStartedPulling="2026-02-15 20:39:53.810418452 +0000 UTC m=+1401.676434085" lastFinishedPulling="2026-02-15 20:40:05.669018413 +0000 UTC m=+1413.535034056" observedRunningTime="2026-02-15 20:40:06.727449412 +0000 UTC m=+1414.593465035" watchObservedRunningTime="2026-02-15 20:40:06.736049571 +0000 UTC m=+1414.602065204" Feb 15 20:40:14 crc kubenswrapper[4735]: I0215 20:40:14.829608 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nbxhf" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="registry-server" probeResult="failure" output=< Feb 15 20:40:14 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:40:14 crc kubenswrapper[4735]: > Feb 15 20:40:18 crc kubenswrapper[4735]: I0215 20:40:18.200842 4735 generic.go:334] "Generic (PLEG): container finished" podID="e31b2a4c-8571-4a0e-868f-f86b757785ac" containerID="71989ec6b9c1f2ab3e1a00951d7ef430245a06736e62846f7266b8dbd7b3e77a" exitCode=0 Feb 15 20:40:18 crc kubenswrapper[4735]: I0215 20:40:18.200912 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" event={"ID":"e31b2a4c-8571-4a0e-868f-f86b757785ac","Type":"ContainerDied","Data":"71989ec6b9c1f2ab3e1a00951d7ef430245a06736e62846f7266b8dbd7b3e77a"} Feb 15 20:40:19 crc kubenswrapper[4735]: I0215 20:40:19.771045 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:40:19 crc kubenswrapper[4735]: I0215 20:40:19.943832 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-ssh-key-openstack-edpm-ipam\") pod \"e31b2a4c-8571-4a0e-868f-f86b757785ac\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " Feb 15 20:40:19 crc kubenswrapper[4735]: I0215 20:40:19.943879 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stdhj\" (UniqueName: \"kubernetes.io/projected/e31b2a4c-8571-4a0e-868f-f86b757785ac-kube-api-access-stdhj\") pod \"e31b2a4c-8571-4a0e-868f-f86b757785ac\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " Feb 15 20:40:19 crc kubenswrapper[4735]: I0215 20:40:19.944008 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-inventory\") pod \"e31b2a4c-8571-4a0e-868f-f86b757785ac\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " Feb 15 20:40:19 crc kubenswrapper[4735]: I0215 20:40:19.944054 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-repo-setup-combined-ca-bundle\") pod \"e31b2a4c-8571-4a0e-868f-f86b757785ac\" (UID: \"e31b2a4c-8571-4a0e-868f-f86b757785ac\") " Feb 15 20:40:19 crc kubenswrapper[4735]: I0215 20:40:19.964255 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e31b2a4c-8571-4a0e-868f-f86b757785ac" (UID: "e31b2a4c-8571-4a0e-868f-f86b757785ac"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:40:19 crc kubenswrapper[4735]: I0215 20:40:19.965678 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31b2a4c-8571-4a0e-868f-f86b757785ac-kube-api-access-stdhj" (OuterVolumeSpecName: "kube-api-access-stdhj") pod "e31b2a4c-8571-4a0e-868f-f86b757785ac" (UID: "e31b2a4c-8571-4a0e-868f-f86b757785ac"). InnerVolumeSpecName "kube-api-access-stdhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:40:19 crc kubenswrapper[4735]: I0215 20:40:19.985323 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-inventory" (OuterVolumeSpecName: "inventory") pod "e31b2a4c-8571-4a0e-868f-f86b757785ac" (UID: "e31b2a4c-8571-4a0e-868f-f86b757785ac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:40:19 crc kubenswrapper[4735]: I0215 20:40:19.985838 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e31b2a4c-8571-4a0e-868f-f86b757785ac" (UID: "e31b2a4c-8571-4a0e-868f-f86b757785ac"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.046866 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.046905 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stdhj\" (UniqueName: \"kubernetes.io/projected/e31b2a4c-8571-4a0e-868f-f86b757785ac-kube-api-access-stdhj\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.046918 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.046931 4735 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31b2a4c-8571-4a0e-868f-f86b757785ac-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.222173 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" event={"ID":"e31b2a4c-8571-4a0e-868f-f86b757785ac","Type":"ContainerDied","Data":"beedb66ff5e1f276cd59355eafe3e5cfad2973819f2221161b101135fb190ea6"} Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.222449 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="beedb66ff5e1f276cd59355eafe3e5cfad2973819f2221161b101135fb190ea6" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.222264 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.331162 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9"] Feb 15 20:40:20 crc kubenswrapper[4735]: E0215 20:40:20.331644 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31b2a4c-8571-4a0e-868f-f86b757785ac" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.331667 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31b2a4c-8571-4a0e-868f-f86b757785ac" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.331901 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31b2a4c-8571-4a0e-868f-f86b757785ac" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.332659 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.339121 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.339327 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.339713 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.339909 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.362081 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9"] Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.453922 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-df8c9\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.454071 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg26g\" (UniqueName: \"kubernetes.io/projected/aad75417-7c89-443b-88ed-aad98d4109b2-kube-api-access-wg26g\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-df8c9\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.454111 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-df8c9\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.555548 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-df8c9\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.555684 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg26g\" (UniqueName: \"kubernetes.io/projected/aad75417-7c89-443b-88ed-aad98d4109b2-kube-api-access-wg26g\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-df8c9\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.555726 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-df8c9\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.559039 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-df8c9\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.561472 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-df8c9\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.573120 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg26g\" (UniqueName: \"kubernetes.io/projected/aad75417-7c89-443b-88ed-aad98d4109b2-kube-api-access-wg26g\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-df8c9\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:20 crc kubenswrapper[4735]: I0215 20:40:20.655922 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:21 crc kubenswrapper[4735]: I0215 20:40:21.221186 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9"] Feb 15 20:40:22 crc kubenswrapper[4735]: I0215 20:40:22.243424 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" event={"ID":"aad75417-7c89-443b-88ed-aad98d4109b2","Type":"ContainerStarted","Data":"e5866e159f6af6c19aaf4a8c1081ba2d5bca969e49610f62a7727203feed3c44"} Feb 15 20:40:22 crc kubenswrapper[4735]: I0215 20:40:22.243784 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" event={"ID":"aad75417-7c89-443b-88ed-aad98d4109b2","Type":"ContainerStarted","Data":"c27dbcf1af52ca8dbce6b7eaa54595f6fac9ec693d4413a6e63b579397d408f0"} Feb 15 20:40:22 crc kubenswrapper[4735]: I0215 20:40:22.260862 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" podStartSLOduration=1.8178847139999998 podStartE2EDuration="2.260843689s" podCreationTimestamp="2026-02-15 20:40:20 +0000 UTC" firstStartedPulling="2026-02-15 20:40:21.228272012 +0000 UTC m=+1429.094287645" lastFinishedPulling="2026-02-15 20:40:21.671230997 +0000 UTC m=+1429.537246620" observedRunningTime="2026-02-15 20:40:22.257758477 +0000 UTC m=+1430.123774110" watchObservedRunningTime="2026-02-15 20:40:22.260843689 +0000 UTC m=+1430.126859312" Feb 15 20:40:23 crc kubenswrapper[4735]: I0215 20:40:23.823671 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:40:23 crc kubenswrapper[4735]: I0215 20:40:23.871614 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:40:24 crc kubenswrapper[4735]: I0215 20:40:24.063602 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nbxhf"] Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.278660 4735 generic.go:334] "Generic (PLEG): container finished" podID="aad75417-7c89-443b-88ed-aad98d4109b2" containerID="e5866e159f6af6c19aaf4a8c1081ba2d5bca969e49610f62a7727203feed3c44" exitCode=0 Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.278807 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" event={"ID":"aad75417-7c89-443b-88ed-aad98d4109b2","Type":"ContainerDied","Data":"e5866e159f6af6c19aaf4a8c1081ba2d5bca969e49610f62a7727203feed3c44"} Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.279484 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nbxhf" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="registry-server" containerID="cri-o://0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679" gracePeriod=2 Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.813364 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.861279 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-utilities\") pod \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.861528 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-catalog-content\") pod \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.861569 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mwm9\" (UniqueName: \"kubernetes.io/projected/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-kube-api-access-9mwm9\") pod \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\" (UID: \"fef0cbc1-3bb0-49e2-a5d0-4680f869c150\") " Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.862664 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-utilities" (OuterVolumeSpecName: "utilities") pod "fef0cbc1-3bb0-49e2-a5d0-4680f869c150" (UID: "fef0cbc1-3bb0-49e2-a5d0-4680f869c150"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.875130 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-kube-api-access-9mwm9" (OuterVolumeSpecName: "kube-api-access-9mwm9") pod "fef0cbc1-3bb0-49e2-a5d0-4680f869c150" (UID: "fef0cbc1-3bb0-49e2-a5d0-4680f869c150"). InnerVolumeSpecName "kube-api-access-9mwm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.964389 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mwm9\" (UniqueName: \"kubernetes.io/projected/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-kube-api-access-9mwm9\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:25 crc kubenswrapper[4735]: I0215 20:40:25.964425 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.009877 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fef0cbc1-3bb0-49e2-a5d0-4680f869c150" (UID: "fef0cbc1-3bb0-49e2-a5d0-4680f869c150"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.067299 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef0cbc1-3bb0-49e2-a5d0-4680f869c150-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.291835 4735 generic.go:334] "Generic (PLEG): container finished" podID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerID="0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679" exitCode=0 Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.291924 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nbxhf" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.292013 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbxhf" event={"ID":"fef0cbc1-3bb0-49e2-a5d0-4680f869c150","Type":"ContainerDied","Data":"0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679"} Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.292057 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbxhf" event={"ID":"fef0cbc1-3bb0-49e2-a5d0-4680f869c150","Type":"ContainerDied","Data":"8a41def63a801d9d985ae6f99d64a623a0c9e0d566bb47f12e849ade5fbf4363"} Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.292085 4735 scope.go:117] "RemoveContainer" containerID="0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.333415 4735 scope.go:117] "RemoveContainer" containerID="376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.348013 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nbxhf"] Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.359538 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nbxhf"] Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.398189 4735 scope.go:117] "RemoveContainer" containerID="770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.448660 4735 scope.go:117] "RemoveContainer" containerID="0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679" Feb 15 20:40:26 crc kubenswrapper[4735]: E0215 20:40:26.451076 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679\": container with ID starting with 0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679 not found: ID does not exist" containerID="0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.451123 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679"} err="failed to get container status \"0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679\": rpc error: code = NotFound desc = could not find container \"0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679\": container with ID starting with 0305786d0fcd4787e031551561b5f498ab4d65e5d3144616764ee8b2583c5679 not found: ID does not exist" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.451148 4735 scope.go:117] "RemoveContainer" containerID="376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b" Feb 15 20:40:26 crc kubenswrapper[4735]: E0215 20:40:26.451426 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b\": container with ID starting with 376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b not found: ID does not exist" containerID="376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.451461 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b"} err="failed to get container status \"376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b\": rpc error: code = NotFound desc = could not find container \"376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b\": container with ID starting with 376e1e61fe70ce20f66e7ecaf494eba5531d49bebfe3d21ba87873ad28c1e37b not found: ID does not exist" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.451486 4735 scope.go:117] "RemoveContainer" containerID="770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502" Feb 15 20:40:26 crc kubenswrapper[4735]: E0215 20:40:26.452388 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502\": container with ID starting with 770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502 not found: ID does not exist" containerID="770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.452406 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502"} err="failed to get container status \"770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502\": rpc error: code = NotFound desc = could not find container \"770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502\": container with ID starting with 770dcb22fe05436b64429407bb82d3fe521991a9033d6af73329accb644d7502 not found: ID does not exist" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.735843 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.783453 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-ssh-key-openstack-edpm-ipam\") pod \"aad75417-7c89-443b-88ed-aad98d4109b2\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.783541 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg26g\" (UniqueName: \"kubernetes.io/projected/aad75417-7c89-443b-88ed-aad98d4109b2-kube-api-access-wg26g\") pod \"aad75417-7c89-443b-88ed-aad98d4109b2\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.783622 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-inventory\") pod \"aad75417-7c89-443b-88ed-aad98d4109b2\" (UID: \"aad75417-7c89-443b-88ed-aad98d4109b2\") " Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.789901 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad75417-7c89-443b-88ed-aad98d4109b2-kube-api-access-wg26g" (OuterVolumeSpecName: "kube-api-access-wg26g") pod "aad75417-7c89-443b-88ed-aad98d4109b2" (UID: "aad75417-7c89-443b-88ed-aad98d4109b2"). InnerVolumeSpecName "kube-api-access-wg26g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.818655 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-inventory" (OuterVolumeSpecName: "inventory") pod "aad75417-7c89-443b-88ed-aad98d4109b2" (UID: "aad75417-7c89-443b-88ed-aad98d4109b2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.830488 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "aad75417-7c89-443b-88ed-aad98d4109b2" (UID: "aad75417-7c89-443b-88ed-aad98d4109b2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.885509 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.885538 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg26g\" (UniqueName: \"kubernetes.io/projected/aad75417-7c89-443b-88ed-aad98d4109b2-kube-api-access-wg26g\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.885548 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad75417-7c89-443b-88ed-aad98d4109b2-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:40:26 crc kubenswrapper[4735]: I0215 20:40:26.895752 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" path="/var/lib/kubelet/pods/fef0cbc1-3bb0-49e2-a5d0-4680f869c150/volumes" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.302692 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.302693 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-df8c9" event={"ID":"aad75417-7c89-443b-88ed-aad98d4109b2","Type":"ContainerDied","Data":"c27dbcf1af52ca8dbce6b7eaa54595f6fac9ec693d4413a6e63b579397d408f0"} Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.302988 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c27dbcf1af52ca8dbce6b7eaa54595f6fac9ec693d4413a6e63b579397d408f0" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.413885 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr"] Feb 15 20:40:27 crc kubenswrapper[4735]: E0215 20:40:27.414323 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="registry-server" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.414344 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="registry-server" Feb 15 20:40:27 crc kubenswrapper[4735]: E0215 20:40:27.414364 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="extract-content" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.414373 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="extract-content" Feb 15 20:40:27 crc kubenswrapper[4735]: E0215 20:40:27.414403 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad75417-7c89-443b-88ed-aad98d4109b2" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.414414 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad75417-7c89-443b-88ed-aad98d4109b2" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 15 20:40:27 crc kubenswrapper[4735]: E0215 20:40:27.414430 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="extract-utilities" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.414438 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="extract-utilities" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.414667 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad75417-7c89-443b-88ed-aad98d4109b2" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.414692 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="fef0cbc1-3bb0-49e2-a5d0-4680f869c150" containerName="registry-server" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.415364 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.417787 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.418047 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.418098 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.421333 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.426804 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr"] Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.495664 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8hkx\" (UniqueName: \"kubernetes.io/projected/b5df6771-b0ed-45d4-aeac-219bb79caf7a-kube-api-access-g8hkx\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.496014 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.496175 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.496568 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.598327 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.598432 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.598488 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8hkx\" (UniqueName: \"kubernetes.io/projected/b5df6771-b0ed-45d4-aeac-219bb79caf7a-kube-api-access-g8hkx\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.598538 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.603691 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.604620 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.615314 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.622920 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8hkx\" (UniqueName: \"kubernetes.io/projected/b5df6771-b0ed-45d4-aeac-219bb79caf7a-kube-api-access-g8hkx\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:27 crc kubenswrapper[4735]: I0215 20:40:27.736933 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:40:28 crc kubenswrapper[4735]: I0215 20:40:28.325812 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr"] Feb 15 20:40:29 crc kubenswrapper[4735]: I0215 20:40:29.325426 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" event={"ID":"b5df6771-b0ed-45d4-aeac-219bb79caf7a","Type":"ContainerStarted","Data":"346ea69e65ceae8a01f9e9e3de351a542e2c5a35602200a7cf6cf92b71f8673b"} Feb 15 20:40:29 crc kubenswrapper[4735]: I0215 20:40:29.325991 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" event={"ID":"b5df6771-b0ed-45d4-aeac-219bb79caf7a","Type":"ContainerStarted","Data":"1320743dd6380b03be09d460931986ace64c929929dad86932478944cc9847b9"} Feb 15 20:40:29 crc kubenswrapper[4735]: I0215 20:40:29.347258 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" podStartSLOduration=1.888941361 podStartE2EDuration="2.347242224s" podCreationTimestamp="2026-02-15 20:40:27 +0000 UTC" firstStartedPulling="2026-02-15 20:40:28.330619603 +0000 UTC m=+1436.196635226" lastFinishedPulling="2026-02-15 20:40:28.788920426 +0000 UTC m=+1436.654936089" observedRunningTime="2026-02-15 20:40:29.342382013 +0000 UTC m=+1437.208397636" watchObservedRunningTime="2026-02-15 20:40:29.347242224 +0000 UTC m=+1437.213257847" Feb 15 20:40:36 crc kubenswrapper[4735]: I0215 20:40:36.484373 4735 scope.go:117] "RemoveContainer" containerID="1bfc08d2407ef98f0c67d479ee227ecbe94d6bfa9ac6f4963f0cdc56de584c08" Feb 15 20:40:36 crc kubenswrapper[4735]: I0215 20:40:36.517556 4735 scope.go:117] "RemoveContainer" containerID="400b8b10f6b7cc5e6032136e65cbc7e66eaf5ed5ded8ec5da99f57d7efae3bb4" Feb 15 20:40:36 crc kubenswrapper[4735]: I0215 20:40:36.567228 4735 scope.go:117] "RemoveContainer" containerID="279b8c3fba818d33c3c3e9ea5ef7e22c36679f2add8fd9f8c9f2d3aae9857205" Feb 15 20:41:36 crc kubenswrapper[4735]: I0215 20:41:36.716793 4735 scope.go:117] "RemoveContainer" containerID="5d80cb6d2dc2ccf6ac05974c0b49e10cfa35444bc235b804bde522b9f6bb18fb" Feb 15 20:41:36 crc kubenswrapper[4735]: I0215 20:41:36.744408 4735 scope.go:117] "RemoveContainer" containerID="e937066142a5010c8285f2c4f18da54ca49cea00b016e120327ee7c2d11c3bfd" Feb 15 20:42:19 crc kubenswrapper[4735]: I0215 20:42:19.679880 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:42:19 crc kubenswrapper[4735]: I0215 20:42:19.680448 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.794345 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2jb7d"] Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.797228 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.839995 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jb7d"] Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.888252 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnkpn\" (UniqueName: \"kubernetes.io/projected/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-kube-api-access-xnkpn\") pod \"redhat-marketplace-2jb7d\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.888342 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-catalog-content\") pod \"redhat-marketplace-2jb7d\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.888383 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-utilities\") pod \"redhat-marketplace-2jb7d\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.990147 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-utilities\") pod \"redhat-marketplace-2jb7d\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.990341 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnkpn\" (UniqueName: \"kubernetes.io/projected/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-kube-api-access-xnkpn\") pod \"redhat-marketplace-2jb7d\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.990445 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-catalog-content\") pod \"redhat-marketplace-2jb7d\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.990580 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-utilities\") pod \"redhat-marketplace-2jb7d\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:30 crc kubenswrapper[4735]: I0215 20:42:30.990730 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-catalog-content\") pod \"redhat-marketplace-2jb7d\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:31 crc kubenswrapper[4735]: I0215 20:42:31.010704 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnkpn\" (UniqueName: \"kubernetes.io/projected/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-kube-api-access-xnkpn\") pod \"redhat-marketplace-2jb7d\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:31 crc kubenswrapper[4735]: I0215 20:42:31.118877 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:31 crc kubenswrapper[4735]: I0215 20:42:31.712084 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jb7d"] Feb 15 20:42:31 crc kubenswrapper[4735]: I0215 20:42:31.769490 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jb7d" event={"ID":"ce5bb53d-f72e-4195-b2d5-d1812fa67da7","Type":"ContainerStarted","Data":"e8b3f29f2228330159fc551395f04280d44381f5da0b8226a5b967f054df035b"} Feb 15 20:42:32 crc kubenswrapper[4735]: I0215 20:42:32.786550 4735 generic.go:334] "Generic (PLEG): container finished" podID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerID="87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47" exitCode=0 Feb 15 20:42:32 crc kubenswrapper[4735]: I0215 20:42:32.786910 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jb7d" event={"ID":"ce5bb53d-f72e-4195-b2d5-d1812fa67da7","Type":"ContainerDied","Data":"87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47"} Feb 15 20:42:34 crc kubenswrapper[4735]: I0215 20:42:34.807438 4735 generic.go:334] "Generic (PLEG): container finished" podID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerID="9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a" exitCode=0 Feb 15 20:42:34 crc kubenswrapper[4735]: I0215 20:42:34.807552 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jb7d" event={"ID":"ce5bb53d-f72e-4195-b2d5-d1812fa67da7","Type":"ContainerDied","Data":"9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a"} Feb 15 20:42:35 crc kubenswrapper[4735]: I0215 20:42:35.817070 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jb7d" event={"ID":"ce5bb53d-f72e-4195-b2d5-d1812fa67da7","Type":"ContainerStarted","Data":"1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0"} Feb 15 20:42:35 crc kubenswrapper[4735]: I0215 20:42:35.843861 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2jb7d" podStartSLOduration=3.207942333 podStartE2EDuration="5.843843256s" podCreationTimestamp="2026-02-15 20:42:30 +0000 UTC" firstStartedPulling="2026-02-15 20:42:32.78975461 +0000 UTC m=+1560.655770233" lastFinishedPulling="2026-02-15 20:42:35.425655523 +0000 UTC m=+1563.291671156" observedRunningTime="2026-02-15 20:42:35.837525666 +0000 UTC m=+1563.703541289" watchObservedRunningTime="2026-02-15 20:42:35.843843256 +0000 UTC m=+1563.709858879" Feb 15 20:42:41 crc kubenswrapper[4735]: I0215 20:42:41.119578 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:41 crc kubenswrapper[4735]: I0215 20:42:41.120009 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:41 crc kubenswrapper[4735]: I0215 20:42:41.183771 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:41 crc kubenswrapper[4735]: I0215 20:42:41.954399 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:42 crc kubenswrapper[4735]: I0215 20:42:42.008021 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jb7d"] Feb 15 20:42:43 crc kubenswrapper[4735]: I0215 20:42:43.895334 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2jb7d" podUID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerName="registry-server" containerID="cri-o://1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0" gracePeriod=2 Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.390889 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.455822 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-catalog-content\") pod \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.455980 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnkpn\" (UniqueName: \"kubernetes.io/projected/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-kube-api-access-xnkpn\") pod \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.456042 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-utilities\") pod \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\" (UID: \"ce5bb53d-f72e-4195-b2d5-d1812fa67da7\") " Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.457198 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-utilities" (OuterVolumeSpecName: "utilities") pod "ce5bb53d-f72e-4195-b2d5-d1812fa67da7" (UID: "ce5bb53d-f72e-4195-b2d5-d1812fa67da7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.479064 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-kube-api-access-xnkpn" (OuterVolumeSpecName: "kube-api-access-xnkpn") pod "ce5bb53d-f72e-4195-b2d5-d1812fa67da7" (UID: "ce5bb53d-f72e-4195-b2d5-d1812fa67da7"). InnerVolumeSpecName "kube-api-access-xnkpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.499430 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce5bb53d-f72e-4195-b2d5-d1812fa67da7" (UID: "ce5bb53d-f72e-4195-b2d5-d1812fa67da7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.573017 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnkpn\" (UniqueName: \"kubernetes.io/projected/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-kube-api-access-xnkpn\") on node \"crc\" DevicePath \"\"" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.573072 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.573094 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5bb53d-f72e-4195-b2d5-d1812fa67da7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.910830 4735 generic.go:334] "Generic (PLEG): container finished" podID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerID="1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0" exitCode=0 Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.910891 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jb7d" event={"ID":"ce5bb53d-f72e-4195-b2d5-d1812fa67da7","Type":"ContainerDied","Data":"1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0"} Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.910921 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jb7d" event={"ID":"ce5bb53d-f72e-4195-b2d5-d1812fa67da7","Type":"ContainerDied","Data":"e8b3f29f2228330159fc551395f04280d44381f5da0b8226a5b967f054df035b"} Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.910927 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2jb7d" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.910956 4735 scope.go:117] "RemoveContainer" containerID="1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.937463 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jb7d"] Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.943786 4735 scope.go:117] "RemoveContainer" containerID="9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a" Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.954306 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jb7d"] Feb 15 20:42:44 crc kubenswrapper[4735]: I0215 20:42:44.967098 4735 scope.go:117] "RemoveContainer" containerID="87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47" Feb 15 20:42:45 crc kubenswrapper[4735]: I0215 20:42:45.009781 4735 scope.go:117] "RemoveContainer" containerID="1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0" Feb 15 20:42:45 crc kubenswrapper[4735]: E0215 20:42:45.010411 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0\": container with ID starting with 1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0 not found: ID does not exist" containerID="1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0" Feb 15 20:42:45 crc kubenswrapper[4735]: I0215 20:42:45.010498 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0"} err="failed to get container status \"1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0\": rpc error: code = NotFound desc = could not find container \"1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0\": container with ID starting with 1597079bc03cbaa36bb14ece2c3fe003dd46ad1be04a83e85eee817a3fff44b0 not found: ID does not exist" Feb 15 20:42:45 crc kubenswrapper[4735]: I0215 20:42:45.010536 4735 scope.go:117] "RemoveContainer" containerID="9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a" Feb 15 20:42:45 crc kubenswrapper[4735]: E0215 20:42:45.010798 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a\": container with ID starting with 9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a not found: ID does not exist" containerID="9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a" Feb 15 20:42:45 crc kubenswrapper[4735]: I0215 20:42:45.010820 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a"} err="failed to get container status \"9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a\": rpc error: code = NotFound desc = could not find container \"9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a\": container with ID starting with 9d90a515db499ae627acc7954afedacf4289b427c9f811db59b2e41470876a1a not found: ID does not exist" Feb 15 20:42:45 crc kubenswrapper[4735]: I0215 20:42:45.010835 4735 scope.go:117] "RemoveContainer" containerID="87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47" Feb 15 20:42:45 crc kubenswrapper[4735]: E0215 20:42:45.011414 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47\": container with ID starting with 87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47 not found: ID does not exist" containerID="87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47" Feb 15 20:42:45 crc kubenswrapper[4735]: I0215 20:42:45.011440 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47"} err="failed to get container status \"87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47\": rpc error: code = NotFound desc = could not find container \"87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47\": container with ID starting with 87735b7bc0149f7c22d31b28fba827e5569ba1cba4fec7d64dd0733b3ae68f47 not found: ID does not exist" Feb 15 20:42:46 crc kubenswrapper[4735]: I0215 20:42:46.899328 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" path="/var/lib/kubelet/pods/ce5bb53d-f72e-4195-b2d5-d1812fa67da7/volumes" Feb 15 20:42:49 crc kubenswrapper[4735]: I0215 20:42:49.680089 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:42:49 crc kubenswrapper[4735]: I0215 20:42:49.680445 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.526754 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zwp92"] Feb 15 20:43:03 crc kubenswrapper[4735]: E0215 20:43:03.527645 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerName="extract-utilities" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.527657 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerName="extract-utilities" Feb 15 20:43:03 crc kubenswrapper[4735]: E0215 20:43:03.527668 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerName="extract-content" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.527674 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerName="extract-content" Feb 15 20:43:03 crc kubenswrapper[4735]: E0215 20:43:03.527685 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerName="registry-server" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.527691 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerName="registry-server" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.527894 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5bb53d-f72e-4195-b2d5-d1812fa67da7" containerName="registry-server" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.529189 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.537844 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwp92"] Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.603246 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-utilities\") pod \"community-operators-zwp92\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.603339 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-catalog-content\") pod \"community-operators-zwp92\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.603398 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2slkl\" (UniqueName: \"kubernetes.io/projected/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-kube-api-access-2slkl\") pod \"community-operators-zwp92\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.704769 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-catalog-content\") pod \"community-operators-zwp92\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.704836 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2slkl\" (UniqueName: \"kubernetes.io/projected/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-kube-api-access-2slkl\") pod \"community-operators-zwp92\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.704935 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-utilities\") pod \"community-operators-zwp92\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.705232 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-catalog-content\") pod \"community-operators-zwp92\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.705338 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-utilities\") pod \"community-operators-zwp92\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.725357 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2slkl\" (UniqueName: \"kubernetes.io/projected/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-kube-api-access-2slkl\") pod \"community-operators-zwp92\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:03 crc kubenswrapper[4735]: I0215 20:43:03.909989 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:04 crc kubenswrapper[4735]: I0215 20:43:04.390325 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwp92"] Feb 15 20:43:05 crc kubenswrapper[4735]: I0215 20:43:05.126587 4735 generic.go:334] "Generic (PLEG): container finished" podID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerID="c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11" exitCode=0 Feb 15 20:43:05 crc kubenswrapper[4735]: I0215 20:43:05.126667 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwp92" event={"ID":"0b712d75-cff8-43e6-8cf0-4427b81b6cf0","Type":"ContainerDied","Data":"c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11"} Feb 15 20:43:05 crc kubenswrapper[4735]: I0215 20:43:05.126852 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwp92" event={"ID":"0b712d75-cff8-43e6-8cf0-4427b81b6cf0","Type":"ContainerStarted","Data":"67d07dac0d1ae00d1071b682ce21baafbb1b08b8b2c070e64ea5586159aaadef"} Feb 15 20:43:05 crc kubenswrapper[4735]: I0215 20:43:05.129094 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 20:43:06 crc kubenswrapper[4735]: I0215 20:43:06.137213 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwp92" event={"ID":"0b712d75-cff8-43e6-8cf0-4427b81b6cf0","Type":"ContainerStarted","Data":"c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93"} Feb 15 20:43:08 crc kubenswrapper[4735]: I0215 20:43:08.199420 4735 generic.go:334] "Generic (PLEG): container finished" podID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerID="c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93" exitCode=0 Feb 15 20:43:08 crc kubenswrapper[4735]: I0215 20:43:08.199930 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwp92" event={"ID":"0b712d75-cff8-43e6-8cf0-4427b81b6cf0","Type":"ContainerDied","Data":"c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93"} Feb 15 20:43:09 crc kubenswrapper[4735]: I0215 20:43:09.227531 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwp92" event={"ID":"0b712d75-cff8-43e6-8cf0-4427b81b6cf0","Type":"ContainerStarted","Data":"c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c"} Feb 15 20:43:09 crc kubenswrapper[4735]: I0215 20:43:09.256143 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zwp92" podStartSLOduration=2.776624836 podStartE2EDuration="6.256120636s" podCreationTimestamp="2026-02-15 20:43:03 +0000 UTC" firstStartedPulling="2026-02-15 20:43:05.128845169 +0000 UTC m=+1592.994860782" lastFinishedPulling="2026-02-15 20:43:08.608340959 +0000 UTC m=+1596.474356582" observedRunningTime="2026-02-15 20:43:09.252693824 +0000 UTC m=+1597.118709457" watchObservedRunningTime="2026-02-15 20:43:09.256120636 +0000 UTC m=+1597.122136259" Feb 15 20:43:13 crc kubenswrapper[4735]: I0215 20:43:13.911155 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:13 crc kubenswrapper[4735]: I0215 20:43:13.912671 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:13 crc kubenswrapper[4735]: I0215 20:43:13.971790 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:14 crc kubenswrapper[4735]: I0215 20:43:14.331248 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:14 crc kubenswrapper[4735]: I0215 20:43:14.379656 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwp92"] Feb 15 20:43:16 crc kubenswrapper[4735]: I0215 20:43:16.352155 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zwp92" podUID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerName="registry-server" containerID="cri-o://c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c" gracePeriod=2 Feb 15 20:43:16 crc kubenswrapper[4735]: I0215 20:43:16.898287 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:16 crc kubenswrapper[4735]: I0215 20:43:16.987798 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2slkl\" (UniqueName: \"kubernetes.io/projected/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-kube-api-access-2slkl\") pod \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " Feb 15 20:43:16 crc kubenswrapper[4735]: I0215 20:43:16.988004 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-catalog-content\") pod \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " Feb 15 20:43:16 crc kubenswrapper[4735]: I0215 20:43:16.988066 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-utilities\") pod \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\" (UID: \"0b712d75-cff8-43e6-8cf0-4427b81b6cf0\") " Feb 15 20:43:16 crc kubenswrapper[4735]: I0215 20:43:16.989788 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-utilities" (OuterVolumeSpecName: "utilities") pod "0b712d75-cff8-43e6-8cf0-4427b81b6cf0" (UID: "0b712d75-cff8-43e6-8cf0-4427b81b6cf0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.007721 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-kube-api-access-2slkl" (OuterVolumeSpecName: "kube-api-access-2slkl") pod "0b712d75-cff8-43e6-8cf0-4427b81b6cf0" (UID: "0b712d75-cff8-43e6-8cf0-4427b81b6cf0"). InnerVolumeSpecName "kube-api-access-2slkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.052975 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b712d75-cff8-43e6-8cf0-4427b81b6cf0" (UID: "0b712d75-cff8-43e6-8cf0-4427b81b6cf0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.089326 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2slkl\" (UniqueName: \"kubernetes.io/projected/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-kube-api-access-2slkl\") on node \"crc\" DevicePath \"\"" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.089359 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.089368 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b712d75-cff8-43e6-8cf0-4427b81b6cf0-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.363701 4735 generic.go:334] "Generic (PLEG): container finished" podID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerID="c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c" exitCode=0 Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.363892 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwp92" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.363924 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwp92" event={"ID":"0b712d75-cff8-43e6-8cf0-4427b81b6cf0","Type":"ContainerDied","Data":"c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c"} Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.365068 4735 scope.go:117] "RemoveContainer" containerID="c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.365072 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwp92" event={"ID":"0b712d75-cff8-43e6-8cf0-4427b81b6cf0","Type":"ContainerDied","Data":"67d07dac0d1ae00d1071b682ce21baafbb1b08b8b2c070e64ea5586159aaadef"} Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.398591 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwp92"] Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.409605 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zwp92"] Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.411402 4735 scope.go:117] "RemoveContainer" containerID="c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.436345 4735 scope.go:117] "RemoveContainer" containerID="c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.487176 4735 scope.go:117] "RemoveContainer" containerID="c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c" Feb 15 20:43:17 crc kubenswrapper[4735]: E0215 20:43:17.487628 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c\": container with ID starting with c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c not found: ID does not exist" containerID="c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.487690 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c"} err="failed to get container status \"c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c\": rpc error: code = NotFound desc = could not find container \"c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c\": container with ID starting with c2b98a50c4ade32c70d9e6ba31a233460bb13baffdac9607699ce5f645d1900c not found: ID does not exist" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.487722 4735 scope.go:117] "RemoveContainer" containerID="c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93" Feb 15 20:43:17 crc kubenswrapper[4735]: E0215 20:43:17.488138 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93\": container with ID starting with c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93 not found: ID does not exist" containerID="c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.488171 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93"} err="failed to get container status \"c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93\": rpc error: code = NotFound desc = could not find container \"c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93\": container with ID starting with c83d9753d9ee181fb6522119c91eb7e322e26f3f3a6bd838723e9394952e9e93 not found: ID does not exist" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.488192 4735 scope.go:117] "RemoveContainer" containerID="c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11" Feb 15 20:43:17 crc kubenswrapper[4735]: E0215 20:43:17.488666 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11\": container with ID starting with c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11 not found: ID does not exist" containerID="c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11" Feb 15 20:43:17 crc kubenswrapper[4735]: I0215 20:43:17.488688 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11"} err="failed to get container status \"c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11\": rpc error: code = NotFound desc = could not find container \"c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11\": container with ID starting with c76b761b073f1282c0c57f1486a426fe5674eb44bdf0cabba987bf30f1bf7b11 not found: ID does not exist" Feb 15 20:43:18 crc kubenswrapper[4735]: I0215 20:43:18.900534 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" path="/var/lib/kubelet/pods/0b712d75-cff8-43e6-8cf0-4427b81b6cf0/volumes" Feb 15 20:43:19 crc kubenswrapper[4735]: I0215 20:43:19.680080 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:43:19 crc kubenswrapper[4735]: I0215 20:43:19.680135 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:43:19 crc kubenswrapper[4735]: I0215 20:43:19.680174 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:43:19 crc kubenswrapper[4735]: I0215 20:43:19.680845 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:43:19 crc kubenswrapper[4735]: I0215 20:43:19.680899 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" gracePeriod=600 Feb 15 20:43:19 crc kubenswrapper[4735]: E0215 20:43:19.798328 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:43:20 crc kubenswrapper[4735]: I0215 20:43:20.397109 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" exitCode=0 Feb 15 20:43:20 crc kubenswrapper[4735]: I0215 20:43:20.397187 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e"} Feb 15 20:43:20 crc kubenswrapper[4735]: I0215 20:43:20.397475 4735 scope.go:117] "RemoveContainer" containerID="97239e5cb100e71e3f44686a5d1c9c6d1a3b14c88d00878cdd46ec7dbf229a42" Feb 15 20:43:20 crc kubenswrapper[4735]: I0215 20:43:20.398084 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:43:20 crc kubenswrapper[4735]: E0215 20:43:20.398355 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:43:30 crc kubenswrapper[4735]: I0215 20:43:30.887068 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:43:30 crc kubenswrapper[4735]: E0215 20:43:30.887938 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:43:36 crc kubenswrapper[4735]: I0215 20:43:36.864805 4735 scope.go:117] "RemoveContainer" containerID="67b53f3ca82355aac170c86227c023b886165282b3769c53e9f598b06b47689a" Feb 15 20:43:42 crc kubenswrapper[4735]: I0215 20:43:42.609379 4735 generic.go:334] "Generic (PLEG): container finished" podID="b5df6771-b0ed-45d4-aeac-219bb79caf7a" containerID="346ea69e65ceae8a01f9e9e3de351a542e2c5a35602200a7cf6cf92b71f8673b" exitCode=0 Feb 15 20:43:42 crc kubenswrapper[4735]: I0215 20:43:42.609932 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" event={"ID":"b5df6771-b0ed-45d4-aeac-219bb79caf7a","Type":"ContainerDied","Data":"346ea69e65ceae8a01f9e9e3de351a542e2c5a35602200a7cf6cf92b71f8673b"} Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.084693 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.243366 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8hkx\" (UniqueName: \"kubernetes.io/projected/b5df6771-b0ed-45d4-aeac-219bb79caf7a-kube-api-access-g8hkx\") pod \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.243511 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-inventory\") pod \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.243584 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-bootstrap-combined-ca-bundle\") pod \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.243708 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-ssh-key-openstack-edpm-ipam\") pod \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\" (UID: \"b5df6771-b0ed-45d4-aeac-219bb79caf7a\") " Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.249437 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b5df6771-b0ed-45d4-aeac-219bb79caf7a" (UID: "b5df6771-b0ed-45d4-aeac-219bb79caf7a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.250357 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5df6771-b0ed-45d4-aeac-219bb79caf7a-kube-api-access-g8hkx" (OuterVolumeSpecName: "kube-api-access-g8hkx") pod "b5df6771-b0ed-45d4-aeac-219bb79caf7a" (UID: "b5df6771-b0ed-45d4-aeac-219bb79caf7a"). InnerVolumeSpecName "kube-api-access-g8hkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.281906 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-inventory" (OuterVolumeSpecName: "inventory") pod "b5df6771-b0ed-45d4-aeac-219bb79caf7a" (UID: "b5df6771-b0ed-45d4-aeac-219bb79caf7a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.282697 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b5df6771-b0ed-45d4-aeac-219bb79caf7a" (UID: "b5df6771-b0ed-45d4-aeac-219bb79caf7a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.347387 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8hkx\" (UniqueName: \"kubernetes.io/projected/b5df6771-b0ed-45d4-aeac-219bb79caf7a-kube-api-access-g8hkx\") on node \"crc\" DevicePath \"\"" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.348261 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.348343 4735 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.348365 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5df6771-b0ed-45d4-aeac-219bb79caf7a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.627554 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" event={"ID":"b5df6771-b0ed-45d4-aeac-219bb79caf7a","Type":"ContainerDied","Data":"1320743dd6380b03be09d460931986ace64c929929dad86932478944cc9847b9"} Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.627603 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1320743dd6380b03be09d460931986ace64c929929dad86932478944cc9847b9" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.628162 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.723104 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg"] Feb 15 20:43:44 crc kubenswrapper[4735]: E0215 20:43:44.723571 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerName="extract-utilities" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.723583 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerName="extract-utilities" Feb 15 20:43:44 crc kubenswrapper[4735]: E0215 20:43:44.723598 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerName="registry-server" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.723604 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerName="registry-server" Feb 15 20:43:44 crc kubenswrapper[4735]: E0215 20:43:44.723618 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerName="extract-content" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.723625 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerName="extract-content" Feb 15 20:43:44 crc kubenswrapper[4735]: E0215 20:43:44.723636 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5df6771-b0ed-45d4-aeac-219bb79caf7a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.723642 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5df6771-b0ed-45d4-aeac-219bb79caf7a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.723835 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5df6771-b0ed-45d4-aeac-219bb79caf7a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.723856 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b712d75-cff8-43e6-8cf0-4427b81b6cf0" containerName="registry-server" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.724521 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.727741 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.728110 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.729203 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.729406 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.742105 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg"] Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.859739 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.859886 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxjx5\" (UniqueName: \"kubernetes.io/projected/a94a8020-82e8-4b37-add7-d9be8fc04c44-kube-api-access-zxjx5\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.860036 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.961887 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.962077 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.962108 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxjx5\" (UniqueName: \"kubernetes.io/projected/a94a8020-82e8-4b37-add7-d9be8fc04c44-kube-api-access-zxjx5\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.977872 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.977936 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:44 crc kubenswrapper[4735]: I0215 20:43:44.982344 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxjx5\" (UniqueName: \"kubernetes.io/projected/a94a8020-82e8-4b37-add7-d9be8fc04c44-kube-api-access-zxjx5\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:45 crc kubenswrapper[4735]: I0215 20:43:45.040696 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:43:45 crc kubenswrapper[4735]: I0215 20:43:45.684620 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg"] Feb 15 20:43:45 crc kubenswrapper[4735]: I0215 20:43:45.887509 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:43:45 crc kubenswrapper[4735]: E0215 20:43:45.887695 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:43:46 crc kubenswrapper[4735]: I0215 20:43:46.647378 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" event={"ID":"a94a8020-82e8-4b37-add7-d9be8fc04c44","Type":"ContainerStarted","Data":"0d5f56c63432d3892281ad14c06672664edf77a3196297eed59124c20d37a140"} Feb 15 20:43:46 crc kubenswrapper[4735]: I0215 20:43:46.647704 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" event={"ID":"a94a8020-82e8-4b37-add7-d9be8fc04c44","Type":"ContainerStarted","Data":"ea7461100be140febcc8fc98e2b892f250caee7e9f763cb44860c0e6abc01a5e"} Feb 15 20:43:46 crc kubenswrapper[4735]: I0215 20:43:46.669375 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" podStartSLOduration=2.255402941 podStartE2EDuration="2.669353359s" podCreationTimestamp="2026-02-15 20:43:44 +0000 UTC" firstStartedPulling="2026-02-15 20:43:45.693647232 +0000 UTC m=+1633.559662855" lastFinishedPulling="2026-02-15 20:43:46.10759765 +0000 UTC m=+1633.973613273" observedRunningTime="2026-02-15 20:43:46.666460192 +0000 UTC m=+1634.532475855" watchObservedRunningTime="2026-02-15 20:43:46.669353359 +0000 UTC m=+1634.535368982" Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.070470 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e7dc-account-create-update-mrg7s"] Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.083382 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5b20-account-create-update-xbszc"] Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.091733 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-rtdvp"] Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.103115 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e7dc-account-create-update-mrg7s"] Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.111024 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-9l2qw"] Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.118716 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5b20-account-create-update-xbszc"] Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.125991 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-rtdvp"] Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.132876 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-9l2qw"] Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.903763 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49b95b5d-8023-49fd-9af3-828e64cbc1fc" path="/var/lib/kubelet/pods/49b95b5d-8023-49fd-9af3-828e64cbc1fc/volumes" Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.906381 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66c606b7-ff0f-48d8-94c6-8beff123f81e" path="/var/lib/kubelet/pods/66c606b7-ff0f-48d8-94c6-8beff123f81e/volumes" Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.908638 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71b4bdc2-22e8-4109-8210-782c0f222045" path="/var/lib/kubelet/pods/71b4bdc2-22e8-4109-8210-782c0f222045/volumes" Feb 15 20:43:50 crc kubenswrapper[4735]: I0215 20:43:50.911334 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4c17912-155f-4926-b2da-ee8bfe43a473" path="/var/lib/kubelet/pods/e4c17912-155f-4926-b2da-ee8bfe43a473/volumes" Feb 15 20:43:53 crc kubenswrapper[4735]: I0215 20:43:53.035371 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-whjxb"] Feb 15 20:43:53 crc kubenswrapper[4735]: I0215 20:43:53.045670 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-4a5c-account-create-update-tq6qj"] Feb 15 20:43:53 crc kubenswrapper[4735]: I0215 20:43:53.054432 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-whjxb"] Feb 15 20:43:53 crc kubenswrapper[4735]: I0215 20:43:53.061339 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-4a5c-account-create-update-tq6qj"] Feb 15 20:43:54 crc kubenswrapper[4735]: I0215 20:43:54.900461 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c860f14-d28c-4d02-8433-15c65ff6e981" path="/var/lib/kubelet/pods/2c860f14-d28c-4d02-8433-15c65ff6e981/volumes" Feb 15 20:43:54 crc kubenswrapper[4735]: I0215 20:43:54.902412 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9" path="/var/lib/kubelet/pods/9296e2d5-b08f-4b7b-9ccd-6cb25d4b6bd9/volumes" Feb 15 20:43:56 crc kubenswrapper[4735]: I0215 20:43:56.887475 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:43:56 crc kubenswrapper[4735]: E0215 20:43:56.888110 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:44:07 crc kubenswrapper[4735]: I0215 20:44:07.886938 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:44:07 crc kubenswrapper[4735]: E0215 20:44:07.887641 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:44:10 crc kubenswrapper[4735]: I0215 20:44:10.055075 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-frck7"] Feb 15 20:44:10 crc kubenswrapper[4735]: I0215 20:44:10.061728 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-frck7"] Feb 15 20:44:10 crc kubenswrapper[4735]: I0215 20:44:10.900000 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8874208-49c3-4e6b-bdb3-ef40de3d9ec7" path="/var/lib/kubelet/pods/c8874208-49c3-4e6b-bdb3-ef40de3d9ec7/volumes" Feb 15 20:44:19 crc kubenswrapper[4735]: I0215 20:44:19.043242 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-djbjm"] Feb 15 20:44:19 crc kubenswrapper[4735]: I0215 20:44:19.054116 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-djbjm"] Feb 15 20:44:20 crc kubenswrapper[4735]: I0215 20:44:20.902696 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be509f30-9a8a-4cd2-8bf7-b8132bd92a97" path="/var/lib/kubelet/pods/be509f30-9a8a-4cd2-8bf7-b8132bd92a97/volumes" Feb 15 20:44:21 crc kubenswrapper[4735]: I0215 20:44:21.887407 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:44:21 crc kubenswrapper[4735]: E0215 20:44:21.887768 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:44:34 crc kubenswrapper[4735]: I0215 20:44:34.887881 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:44:34 crc kubenswrapper[4735]: E0215 20:44:34.888579 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:44:36 crc kubenswrapper[4735]: I0215 20:44:36.068599 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-p2mmq"] Feb 15 20:44:36 crc kubenswrapper[4735]: I0215 20:44:36.084193 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-p2mmq"] Feb 15 20:44:36 crc kubenswrapper[4735]: I0215 20:44:36.896828 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66" path="/var/lib/kubelet/pods/b1c5a4f9-ac3e-4cd6-953a-53b4399cfc66/volumes" Feb 15 20:44:37 crc kubenswrapper[4735]: I0215 20:44:37.111784 4735 scope.go:117] "RemoveContainer" containerID="4b531105bad682db1ce00829f8cf5b929722989c2d49e1610e1aa8140affcbce" Feb 15 20:44:37 crc kubenswrapper[4735]: I0215 20:44:37.155549 4735 scope.go:117] "RemoveContainer" containerID="045ff36bcd7e2a5303e6b80261e9ada2e1904c716782a5889b1317f9b7b7e67f" Feb 15 20:44:37 crc kubenswrapper[4735]: I0215 20:44:37.244885 4735 scope.go:117] "RemoveContainer" containerID="6f82ddd6d3abf184aed0373376012195adbe6090c34df6d80c183d78e380a37f" Feb 15 20:44:37 crc kubenswrapper[4735]: I0215 20:44:37.316163 4735 scope.go:117] "RemoveContainer" containerID="4e51bbb9c8a5e7751e1502d3e29d703524a71a0e0da362fb564c6b50c832f5a9" Feb 15 20:44:37 crc kubenswrapper[4735]: I0215 20:44:37.351098 4735 scope.go:117] "RemoveContainer" containerID="eb4d060cd69cd3d8740ffd482b2646fad6b20b0c95addca72d272e1cc473bf54" Feb 15 20:44:37 crc kubenswrapper[4735]: I0215 20:44:37.381508 4735 scope.go:117] "RemoveContainer" containerID="45da017dd2d568023786d1d5eb1eeff68778cb3d628305a3246665dd9d8bbb37" Feb 15 20:44:37 crc kubenswrapper[4735]: I0215 20:44:37.423315 4735 scope.go:117] "RemoveContainer" containerID="a5b026cedf3008efb62514c2b8b28bcdba0a68446846e0d37243658973316b8b" Feb 15 20:44:37 crc kubenswrapper[4735]: I0215 20:44:37.445852 4735 scope.go:117] "RemoveContainer" containerID="eb389e6c8ab356ba85b72553b28fdd28b0db1572c97b36a3dc3fb08ccaa72bc1" Feb 15 20:44:37 crc kubenswrapper[4735]: I0215 20:44:37.470634 4735 scope.go:117] "RemoveContainer" containerID="90ce3286cf31e407ce65270185e348d898a3df8f4f741d5c84b84148c2904fb3" Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.046780 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-j6mch"] Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.062606 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-d27lr"] Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.074463 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8ea4-account-create-update-sjn94"] Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.084327 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f244-account-create-update-kc8sg"] Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.093813 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-fd5e-account-create-update-m58kx"] Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.102788 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-d27lr"] Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.110214 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-j6mch"] Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.117784 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f244-account-create-update-kc8sg"] Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.128231 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8ea4-account-create-update-sjn94"] Feb 15 20:44:39 crc kubenswrapper[4735]: I0215 20:44:39.133706 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-fd5e-account-create-update-m58kx"] Feb 15 20:44:40 crc kubenswrapper[4735]: I0215 20:44:40.910105 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36b69cc5-60b4-4560-9a60-27db0295a1cc" path="/var/lib/kubelet/pods/36b69cc5-60b4-4560-9a60-27db0295a1cc/volumes" Feb 15 20:44:40 crc kubenswrapper[4735]: I0215 20:44:40.911912 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b66e376f-3330-4024-a409-c6820011d354" path="/var/lib/kubelet/pods/b66e376f-3330-4024-a409-c6820011d354/volumes" Feb 15 20:44:40 crc kubenswrapper[4735]: I0215 20:44:40.913972 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3bc62da-42b0-4f89-907a-64b5d2443fcf" path="/var/lib/kubelet/pods/d3bc62da-42b0-4f89-907a-64b5d2443fcf/volumes" Feb 15 20:44:40 crc kubenswrapper[4735]: I0215 20:44:40.915735 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2aacebd-28b2-43f7-afaa-966bd5bf7ade" path="/var/lib/kubelet/pods/e2aacebd-28b2-43f7-afaa-966bd5bf7ade/volumes" Feb 15 20:44:40 crc kubenswrapper[4735]: I0215 20:44:40.917675 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f82da625-890e-43b7-8dee-7ab76e375b52" path="/var/lib/kubelet/pods/f82da625-890e-43b7-8dee-7ab76e375b52/volumes" Feb 15 20:44:44 crc kubenswrapper[4735]: I0215 20:44:44.026091 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ft5v7"] Feb 15 20:44:44 crc kubenswrapper[4735]: I0215 20:44:44.034711 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ft5v7"] Feb 15 20:44:44 crc kubenswrapper[4735]: I0215 20:44:44.895354 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c3905d6-692c-4791-84b0-58d7f6372642" path="/var/lib/kubelet/pods/5c3905d6-692c-4791-84b0-58d7f6372642/volumes" Feb 15 20:44:48 crc kubenswrapper[4735]: I0215 20:44:48.887925 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:44:48 crc kubenswrapper[4735]: E0215 20:44:48.888483 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.181713 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586"] Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.183667 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.186682 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.192195 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.193964 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586"] Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.279460 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6msd\" (UniqueName: \"kubernetes.io/projected/e0b06b04-ee13-4172-99b2-6c5552365399-kube-api-access-z6msd\") pod \"collect-profiles-29519805-6b586\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.279526 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0b06b04-ee13-4172-99b2-6c5552365399-config-volume\") pod \"collect-profiles-29519805-6b586\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.279550 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0b06b04-ee13-4172-99b2-6c5552365399-secret-volume\") pod \"collect-profiles-29519805-6b586\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.380599 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6msd\" (UniqueName: \"kubernetes.io/projected/e0b06b04-ee13-4172-99b2-6c5552365399-kube-api-access-z6msd\") pod \"collect-profiles-29519805-6b586\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.380650 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0b06b04-ee13-4172-99b2-6c5552365399-config-volume\") pod \"collect-profiles-29519805-6b586\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.380680 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0b06b04-ee13-4172-99b2-6c5552365399-secret-volume\") pod \"collect-profiles-29519805-6b586\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.381728 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0b06b04-ee13-4172-99b2-6c5552365399-config-volume\") pod \"collect-profiles-29519805-6b586\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.388800 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0b06b04-ee13-4172-99b2-6c5552365399-secret-volume\") pod \"collect-profiles-29519805-6b586\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.402021 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6msd\" (UniqueName: \"kubernetes.io/projected/e0b06b04-ee13-4172-99b2-6c5552365399-kube-api-access-z6msd\") pod \"collect-profiles-29519805-6b586\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.540725 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:00 crc kubenswrapper[4735]: I0215 20:45:00.899742 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:45:00 crc kubenswrapper[4735]: E0215 20:45:00.900502 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:45:01 crc kubenswrapper[4735]: I0215 20:45:01.005467 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586"] Feb 15 20:45:01 crc kubenswrapper[4735]: I0215 20:45:01.335384 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" event={"ID":"e0b06b04-ee13-4172-99b2-6c5552365399","Type":"ContainerStarted","Data":"f3f1ee0cfa7e32ca2f5f444fc01874fd51e9fe0b6403263e19089cc8771530d4"} Feb 15 20:45:01 crc kubenswrapper[4735]: I0215 20:45:01.335435 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" event={"ID":"e0b06b04-ee13-4172-99b2-6c5552365399","Type":"ContainerStarted","Data":"22cc9fc5263465c90121f56022f298521156a09b918d2cdb3193745fae3a14d2"} Feb 15 20:45:01 crc kubenswrapper[4735]: I0215 20:45:01.363251 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" podStartSLOduration=1.363217426 podStartE2EDuration="1.363217426s" podCreationTimestamp="2026-02-15 20:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 20:45:01.351462041 +0000 UTC m=+1709.217477714" watchObservedRunningTime="2026-02-15 20:45:01.363217426 +0000 UTC m=+1709.229233099" Feb 15 20:45:02 crc kubenswrapper[4735]: I0215 20:45:02.348826 4735 generic.go:334] "Generic (PLEG): container finished" podID="e0b06b04-ee13-4172-99b2-6c5552365399" containerID="f3f1ee0cfa7e32ca2f5f444fc01874fd51e9fe0b6403263e19089cc8771530d4" exitCode=0 Feb 15 20:45:02 crc kubenswrapper[4735]: I0215 20:45:02.348958 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" event={"ID":"e0b06b04-ee13-4172-99b2-6c5552365399","Type":"ContainerDied","Data":"f3f1ee0cfa7e32ca2f5f444fc01874fd51e9fe0b6403263e19089cc8771530d4"} Feb 15 20:45:03 crc kubenswrapper[4735]: I0215 20:45:03.710223 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:03 crc kubenswrapper[4735]: I0215 20:45:03.852255 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6msd\" (UniqueName: \"kubernetes.io/projected/e0b06b04-ee13-4172-99b2-6c5552365399-kube-api-access-z6msd\") pod \"e0b06b04-ee13-4172-99b2-6c5552365399\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " Feb 15 20:45:03 crc kubenswrapper[4735]: I0215 20:45:03.852293 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0b06b04-ee13-4172-99b2-6c5552365399-secret-volume\") pod \"e0b06b04-ee13-4172-99b2-6c5552365399\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " Feb 15 20:45:03 crc kubenswrapper[4735]: I0215 20:45:03.853026 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0b06b04-ee13-4172-99b2-6c5552365399-config-volume" (OuterVolumeSpecName: "config-volume") pod "e0b06b04-ee13-4172-99b2-6c5552365399" (UID: "e0b06b04-ee13-4172-99b2-6c5552365399"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:45:03 crc kubenswrapper[4735]: I0215 20:45:03.852451 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0b06b04-ee13-4172-99b2-6c5552365399-config-volume\") pod \"e0b06b04-ee13-4172-99b2-6c5552365399\" (UID: \"e0b06b04-ee13-4172-99b2-6c5552365399\") " Feb 15 20:45:03 crc kubenswrapper[4735]: I0215 20:45:03.853526 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0b06b04-ee13-4172-99b2-6c5552365399-config-volume\") on node \"crc\" DevicePath \"\"" Feb 15 20:45:04 crc kubenswrapper[4735]: I0215 20:45:04.220596 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0b06b04-ee13-4172-99b2-6c5552365399-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e0b06b04-ee13-4172-99b2-6c5552365399" (UID: "e0b06b04-ee13-4172-99b2-6c5552365399"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:45:04 crc kubenswrapper[4735]: I0215 20:45:04.265674 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0b06b04-ee13-4172-99b2-6c5552365399-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 15 20:45:04 crc kubenswrapper[4735]: I0215 20:45:04.632208 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0b06b04-ee13-4172-99b2-6c5552365399-kube-api-access-z6msd" (OuterVolumeSpecName: "kube-api-access-z6msd") pod "e0b06b04-ee13-4172-99b2-6c5552365399" (UID: "e0b06b04-ee13-4172-99b2-6c5552365399"). InnerVolumeSpecName "kube-api-access-z6msd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:45:04 crc kubenswrapper[4735]: I0215 20:45:04.670289 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" event={"ID":"e0b06b04-ee13-4172-99b2-6c5552365399","Type":"ContainerDied","Data":"22cc9fc5263465c90121f56022f298521156a09b918d2cdb3193745fae3a14d2"} Feb 15 20:45:04 crc kubenswrapper[4735]: I0215 20:45:04.670341 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22cc9fc5263465c90121f56022f298521156a09b918d2cdb3193745fae3a14d2" Feb 15 20:45:04 crc kubenswrapper[4735]: I0215 20:45:04.670411 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586" Feb 15 20:45:04 crc kubenswrapper[4735]: I0215 20:45:04.719128 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6msd\" (UniqueName: \"kubernetes.io/projected/e0b06b04-ee13-4172-99b2-6c5552365399-kube-api-access-z6msd\") on node \"crc\" DevicePath \"\"" Feb 15 20:45:15 crc kubenswrapper[4735]: I0215 20:45:15.887079 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:45:15 crc kubenswrapper[4735]: E0215 20:45:15.887761 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:45:17 crc kubenswrapper[4735]: I0215 20:45:17.057986 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-hk7jh"] Feb 15 20:45:17 crc kubenswrapper[4735]: I0215 20:45:17.070764 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-hk7jh"] Feb 15 20:45:18 crc kubenswrapper[4735]: I0215 20:45:18.901274 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ae5b3cf-b0c9-4d46-ac26-491df4077db9" path="/var/lib/kubelet/pods/6ae5b3cf-b0c9-4d46-ac26-491df4077db9/volumes" Feb 15 20:45:22 crc kubenswrapper[4735]: I0215 20:45:22.823162 4735 generic.go:334] "Generic (PLEG): container finished" podID="a94a8020-82e8-4b37-add7-d9be8fc04c44" containerID="0d5f56c63432d3892281ad14c06672664edf77a3196297eed59124c20d37a140" exitCode=0 Feb 15 20:45:22 crc kubenswrapper[4735]: I0215 20:45:22.823239 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" event={"ID":"a94a8020-82e8-4b37-add7-d9be8fc04c44","Type":"ContainerDied","Data":"0d5f56c63432d3892281ad14c06672664edf77a3196297eed59124c20d37a140"} Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.307993 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.496758 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-inventory\") pod \"a94a8020-82e8-4b37-add7-d9be8fc04c44\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.496919 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxjx5\" (UniqueName: \"kubernetes.io/projected/a94a8020-82e8-4b37-add7-d9be8fc04c44-kube-api-access-zxjx5\") pod \"a94a8020-82e8-4b37-add7-d9be8fc04c44\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.496980 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-ssh-key-openstack-edpm-ipam\") pod \"a94a8020-82e8-4b37-add7-d9be8fc04c44\" (UID: \"a94a8020-82e8-4b37-add7-d9be8fc04c44\") " Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.517829 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a94a8020-82e8-4b37-add7-d9be8fc04c44-kube-api-access-zxjx5" (OuterVolumeSpecName: "kube-api-access-zxjx5") pod "a94a8020-82e8-4b37-add7-d9be8fc04c44" (UID: "a94a8020-82e8-4b37-add7-d9be8fc04c44"). InnerVolumeSpecName "kube-api-access-zxjx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.528140 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-inventory" (OuterVolumeSpecName: "inventory") pod "a94a8020-82e8-4b37-add7-d9be8fc04c44" (UID: "a94a8020-82e8-4b37-add7-d9be8fc04c44"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.533069 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a94a8020-82e8-4b37-add7-d9be8fc04c44" (UID: "a94a8020-82e8-4b37-add7-d9be8fc04c44"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.598634 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxjx5\" (UniqueName: \"kubernetes.io/projected/a94a8020-82e8-4b37-add7-d9be8fc04c44-kube-api-access-zxjx5\") on node \"crc\" DevicePath \"\"" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.598662 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.598675 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94a8020-82e8-4b37-add7-d9be8fc04c44-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.840636 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" event={"ID":"a94a8020-82e8-4b37-add7-d9be8fc04c44","Type":"ContainerDied","Data":"ea7461100be140febcc8fc98e2b892f250caee7e9f763cb44860c0e6abc01a5e"} Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.840932 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea7461100be140febcc8fc98e2b892f250caee7e9f763cb44860c0e6abc01a5e" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.840700 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.929647 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn"] Feb 15 20:45:24 crc kubenswrapper[4735]: E0215 20:45:24.930006 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0b06b04-ee13-4172-99b2-6c5552365399" containerName="collect-profiles" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.930022 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0b06b04-ee13-4172-99b2-6c5552365399" containerName="collect-profiles" Feb 15 20:45:24 crc kubenswrapper[4735]: E0215 20:45:24.930038 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a94a8020-82e8-4b37-add7-d9be8fc04c44" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.930045 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a94a8020-82e8-4b37-add7-d9be8fc04c44" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.930217 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0b06b04-ee13-4172-99b2-6c5552365399" containerName="collect-profiles" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.930236 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a94a8020-82e8-4b37-add7-d9be8fc04c44" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.930768 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.940939 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.941085 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.941709 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.943153 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:45:24 crc kubenswrapper[4735]: I0215 20:45:24.944654 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn"] Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.106960 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bwldn\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.107058 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bwldn\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.107201 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgkm5\" (UniqueName: \"kubernetes.io/projected/f62f1800-36ef-497f-94af-ca96c2a3f7e9-kube-api-access-wgkm5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bwldn\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.208522 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bwldn\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.208644 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgkm5\" (UniqueName: \"kubernetes.io/projected/f62f1800-36ef-497f-94af-ca96c2a3f7e9-kube-api-access-wgkm5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bwldn\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.208728 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bwldn\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.212805 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bwldn\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.213366 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bwldn\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.234144 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgkm5\" (UniqueName: \"kubernetes.io/projected/f62f1800-36ef-497f-94af-ca96c2a3f7e9-kube-api-access-wgkm5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bwldn\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.246761 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.831703 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn"] Feb 15 20:45:25 crc kubenswrapper[4735]: W0215 20:45:25.835349 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf62f1800_36ef_497f_94af_ca96c2a3f7e9.slice/crio-0ea3e1419833484f002eb65243a0f69c8209473c291dc865a73bbba23c954068 WatchSource:0}: Error finding container 0ea3e1419833484f002eb65243a0f69c8209473c291dc865a73bbba23c954068: Status 404 returned error can't find the container with id 0ea3e1419833484f002eb65243a0f69c8209473c291dc865a73bbba23c954068 Feb 15 20:45:25 crc kubenswrapper[4735]: I0215 20:45:25.863414 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" event={"ID":"f62f1800-36ef-497f-94af-ca96c2a3f7e9","Type":"ContainerStarted","Data":"0ea3e1419833484f002eb65243a0f69c8209473c291dc865a73bbba23c954068"} Feb 15 20:45:26 crc kubenswrapper[4735]: I0215 20:45:26.871973 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" event={"ID":"f62f1800-36ef-497f-94af-ca96c2a3f7e9","Type":"ContainerStarted","Data":"937725ceebfbe00ac8c36c6a0bedfc395ddd236bc9c8bf09c32dde51d1237a42"} Feb 15 20:45:26 crc kubenswrapper[4735]: I0215 20:45:26.895490 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" podStartSLOduration=2.483480037 podStartE2EDuration="2.895468907s" podCreationTimestamp="2026-02-15 20:45:24 +0000 UTC" firstStartedPulling="2026-02-15 20:45:25.837689102 +0000 UTC m=+1733.703704725" lastFinishedPulling="2026-02-15 20:45:26.249677932 +0000 UTC m=+1734.115693595" observedRunningTime="2026-02-15 20:45:26.88769458 +0000 UTC m=+1734.753710223" watchObservedRunningTime="2026-02-15 20:45:26.895468907 +0000 UTC m=+1734.761484550" Feb 15 20:45:27 crc kubenswrapper[4735]: I0215 20:45:27.887868 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:45:27 crc kubenswrapper[4735]: E0215 20:45:27.888183 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:45:34 crc kubenswrapper[4735]: I0215 20:45:34.040380 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-pss7t"] Feb 15 20:45:34 crc kubenswrapper[4735]: I0215 20:45:34.050409 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-pss7t"] Feb 15 20:45:34 crc kubenswrapper[4735]: I0215 20:45:34.059704 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4r5s8"] Feb 15 20:45:34 crc kubenswrapper[4735]: I0215 20:45:34.069570 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ph5qb"] Feb 15 20:45:34 crc kubenswrapper[4735]: I0215 20:45:34.079560 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ph5qb"] Feb 15 20:45:34 crc kubenswrapper[4735]: I0215 20:45:34.086897 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4r5s8"] Feb 15 20:45:34 crc kubenswrapper[4735]: I0215 20:45:34.903309 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="327449fc-a14d-4594-b8a8-c758aeebec34" path="/var/lib/kubelet/pods/327449fc-a14d-4594-b8a8-c758aeebec34/volumes" Feb 15 20:45:34 crc kubenswrapper[4735]: I0215 20:45:34.919573 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76a8192d-7bcd-4f4a-b798-907d67b3d21c" path="/var/lib/kubelet/pods/76a8192d-7bcd-4f4a-b798-907d67b3d21c/volumes" Feb 15 20:45:34 crc kubenswrapper[4735]: I0215 20:45:34.949277 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c924b448-cd94-426c-bc25-ad0efa381777" path="/var/lib/kubelet/pods/c924b448-cd94-426c-bc25-ad0efa381777/volumes" Feb 15 20:45:37 crc kubenswrapper[4735]: I0215 20:45:37.657190 4735 scope.go:117] "RemoveContainer" containerID="230656781d2e7c4b1b9fc47f342546e1fd086719a5ed2672d490bb1ed7e69539" Feb 15 20:45:37 crc kubenswrapper[4735]: I0215 20:45:37.708625 4735 scope.go:117] "RemoveContainer" containerID="a5bc5afa03916a5704a9ec8c371f7ae1f88bd238482b4f6a7ac79a273e6ec903" Feb 15 20:45:37 crc kubenswrapper[4735]: I0215 20:45:37.758596 4735 scope.go:117] "RemoveContainer" containerID="b284e77e8f04e551f96ea3329e76b0508cdfd5b5047ca426e3bb254e9860f63e" Feb 15 20:45:37 crc kubenswrapper[4735]: I0215 20:45:37.808466 4735 scope.go:117] "RemoveContainer" containerID="e4e49c8e14724339528fd40f102fe47b4d97119994894531101083843d09ead7" Feb 15 20:45:37 crc kubenswrapper[4735]: I0215 20:45:37.860228 4735 scope.go:117] "RemoveContainer" containerID="09f9578052a97aa9ebe8cfb14289319207556a64165f83e900c09158d6b673d2" Feb 15 20:45:37 crc kubenswrapper[4735]: I0215 20:45:37.905662 4735 scope.go:117] "RemoveContainer" containerID="de1b3c92adaf1449e8bbe0b59c65e47e6120fb3f52283281fb50db3bb7a8e1aa" Feb 15 20:45:37 crc kubenswrapper[4735]: I0215 20:45:37.963186 4735 scope.go:117] "RemoveContainer" containerID="c2bf42173a97f9145aff4eb6ccff1e5f696023d0eaec1c7d82f9d307e916cc36" Feb 15 20:45:38 crc kubenswrapper[4735]: I0215 20:45:38.012072 4735 scope.go:117] "RemoveContainer" containerID="8fdb6b04f4a90b71fd7b845d209a2074d7f9c23207a44a099f78922ae5354d53" Feb 15 20:45:38 crc kubenswrapper[4735]: I0215 20:45:38.045615 4735 scope.go:117] "RemoveContainer" containerID="410a4036f6910ec4b0f837c8cfff0e0d7d3938ae03838de515d272f46e927a52" Feb 15 20:45:38 crc kubenswrapper[4735]: I0215 20:45:38.085398 4735 scope.go:117] "RemoveContainer" containerID="b646967960384495241fdd563402509ed82e8749048e7b71208e69151ee8e5af" Feb 15 20:45:38 crc kubenswrapper[4735]: I0215 20:45:38.886510 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:45:38 crc kubenswrapper[4735]: E0215 20:45:38.886888 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:45:50 crc kubenswrapper[4735]: I0215 20:45:50.887307 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:45:50 crc kubenswrapper[4735]: E0215 20:45:50.888084 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:45:56 crc kubenswrapper[4735]: I0215 20:45:56.044083 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-6htdh"] Feb 15 20:45:56 crc kubenswrapper[4735]: I0215 20:45:56.051218 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-6htdh"] Feb 15 20:45:56 crc kubenswrapper[4735]: I0215 20:45:56.898342 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec4fe5e3-ca92-4814-8921-62117a037281" path="/var/lib/kubelet/pods/ec4fe5e3-ca92-4814-8921-62117a037281/volumes" Feb 15 20:46:04 crc kubenswrapper[4735]: I0215 20:46:04.888748 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:46:04 crc kubenswrapper[4735]: E0215 20:46:04.890432 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:46:19 crc kubenswrapper[4735]: I0215 20:46:19.886358 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:46:19 crc kubenswrapper[4735]: E0215 20:46:19.887050 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:46:31 crc kubenswrapper[4735]: I0215 20:46:31.887750 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:46:31 crc kubenswrapper[4735]: E0215 20:46:31.888770 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:46:36 crc kubenswrapper[4735]: I0215 20:46:36.043157 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-svbdx"] Feb 15 20:46:36 crc kubenswrapper[4735]: I0215 20:46:36.056613 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-m6ztx"] Feb 15 20:46:36 crc kubenswrapper[4735]: I0215 20:46:36.069517 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-svbdx"] Feb 15 20:46:36 crc kubenswrapper[4735]: I0215 20:46:36.077970 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-m6ztx"] Feb 15 20:46:36 crc kubenswrapper[4735]: I0215 20:46:36.915189 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48189563-f3c7-49a0-b01e-86b25b077fd5" path="/var/lib/kubelet/pods/48189563-f3c7-49a0-b01e-86b25b077fd5/volumes" Feb 15 20:46:36 crc kubenswrapper[4735]: I0215 20:46:36.918208 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8507ec7-d881-4a4a-82dc-3af56d4d53aa" path="/var/lib/kubelet/pods/c8507ec7-d881-4a4a-82dc-3af56d4d53aa/volumes" Feb 15 20:46:37 crc kubenswrapper[4735]: I0215 20:46:37.534052 4735 generic.go:334] "Generic (PLEG): container finished" podID="f62f1800-36ef-497f-94af-ca96c2a3f7e9" containerID="937725ceebfbe00ac8c36c6a0bedfc395ddd236bc9c8bf09c32dde51d1237a42" exitCode=0 Feb 15 20:46:37 crc kubenswrapper[4735]: I0215 20:46:37.534106 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" event={"ID":"f62f1800-36ef-497f-94af-ca96c2a3f7e9","Type":"ContainerDied","Data":"937725ceebfbe00ac8c36c6a0bedfc395ddd236bc9c8bf09c32dde51d1237a42"} Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.035157 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-f66hh"] Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.042786 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-33f9-account-create-update-t72d7"] Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.050385 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-af86-account-create-update-5mp2z"] Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.058154 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-d824-account-create-update-gmj74"] Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.065353 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-33f9-account-create-update-t72d7"] Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.072718 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-f66hh"] Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.078892 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-af86-account-create-update-5mp2z"] Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.085471 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-d824-account-create-update-gmj74"] Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.283137 4735 scope.go:117] "RemoveContainer" containerID="524a5b6f479153a7a9a98fa20f5ed7416bf0165d557b4e405482134294240272" Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.319361 4735 scope.go:117] "RemoveContainer" containerID="3bc53ecea541aa831649411facd7ba4a1417813f89093f199f69eae98e61b2f8" Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.357671 4735 scope.go:117] "RemoveContainer" containerID="0f1548d194a99f08acec43643a3412f38380dea230ed3068fa946c56108d1ea7" Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.902475 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="493558c5-9375-43a2-92e1-9f20470fc681" path="/var/lib/kubelet/pods/493558c5-9375-43a2-92e1-9f20470fc681/volumes" Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.904332 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85484764-8eab-4bbd-8d2f-1726f287ada3" path="/var/lib/kubelet/pods/85484764-8eab-4bbd-8d2f-1726f287ada3/volumes" Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.908663 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f27472c1-f67e-49e6-bd56-a36101f69c64" path="/var/lib/kubelet/pods/f27472c1-f67e-49e6-bd56-a36101f69c64/volumes" Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.909676 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2d80b54-e0f6-415a-9bab-1cff8e208efa" path="/var/lib/kubelet/pods/f2d80b54-e0f6-415a-9bab-1cff8e208efa/volumes" Feb 15 20:46:38 crc kubenswrapper[4735]: I0215 20:46:38.968125 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.101055 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgkm5\" (UniqueName: \"kubernetes.io/projected/f62f1800-36ef-497f-94af-ca96c2a3f7e9-kube-api-access-wgkm5\") pod \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.101093 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-inventory\") pod \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.101171 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-ssh-key-openstack-edpm-ipam\") pod \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\" (UID: \"f62f1800-36ef-497f-94af-ca96c2a3f7e9\") " Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.106833 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f62f1800-36ef-497f-94af-ca96c2a3f7e9-kube-api-access-wgkm5" (OuterVolumeSpecName: "kube-api-access-wgkm5") pod "f62f1800-36ef-497f-94af-ca96c2a3f7e9" (UID: "f62f1800-36ef-497f-94af-ca96c2a3f7e9"). InnerVolumeSpecName "kube-api-access-wgkm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.127261 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-inventory" (OuterVolumeSpecName: "inventory") pod "f62f1800-36ef-497f-94af-ca96c2a3f7e9" (UID: "f62f1800-36ef-497f-94af-ca96c2a3f7e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.149246 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f62f1800-36ef-497f-94af-ca96c2a3f7e9" (UID: "f62f1800-36ef-497f-94af-ca96c2a3f7e9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.202898 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgkm5\" (UniqueName: \"kubernetes.io/projected/f62f1800-36ef-497f-94af-ca96c2a3f7e9-kube-api-access-wgkm5\") on node \"crc\" DevicePath \"\"" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.202941 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.202981 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f62f1800-36ef-497f-94af-ca96c2a3f7e9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.551362 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" event={"ID":"f62f1800-36ef-497f-94af-ca96c2a3f7e9","Type":"ContainerDied","Data":"0ea3e1419833484f002eb65243a0f69c8209473c291dc865a73bbba23c954068"} Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.551790 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ea3e1419833484f002eb65243a0f69c8209473c291dc865a73bbba23c954068" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.551416 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bwldn" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.683538 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch"] Feb 15 20:46:39 crc kubenswrapper[4735]: E0215 20:46:39.684059 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62f1800-36ef-497f-94af-ca96c2a3f7e9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.684076 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62f1800-36ef-497f-94af-ca96c2a3f7e9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.684383 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62f1800-36ef-497f-94af-ca96c2a3f7e9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.689578 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.694405 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.694714 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.694842 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.694974 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.697938 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch"] Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.816981 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7srw\" (UniqueName: \"kubernetes.io/projected/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-kube-api-access-k7srw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.817313 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.817411 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.920052 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.920202 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.921363 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7srw\" (UniqueName: \"kubernetes.io/projected/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-kube-api-access-k7srw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.927822 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.930092 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:39 crc kubenswrapper[4735]: I0215 20:46:39.957802 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7srw\" (UniqueName: \"kubernetes.io/projected/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-kube-api-access-k7srw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:40 crc kubenswrapper[4735]: I0215 20:46:40.006342 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:40 crc kubenswrapper[4735]: I0215 20:46:40.402472 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch"] Feb 15 20:46:40 crc kubenswrapper[4735]: W0215 20:46:40.421582 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c7b61e8_5a6a_4651_bf3b_3f90ae858b40.slice/crio-3c3f66627611dc30b567b5afd8627c12e54e3a6e28d4e5702c54d0a909b346ad WatchSource:0}: Error finding container 3c3f66627611dc30b567b5afd8627c12e54e3a6e28d4e5702c54d0a909b346ad: Status 404 returned error can't find the container with id 3c3f66627611dc30b567b5afd8627c12e54e3a6e28d4e5702c54d0a909b346ad Feb 15 20:46:40 crc kubenswrapper[4735]: I0215 20:46:40.563017 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" event={"ID":"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40","Type":"ContainerStarted","Data":"3c3f66627611dc30b567b5afd8627c12e54e3a6e28d4e5702c54d0a909b346ad"} Feb 15 20:46:41 crc kubenswrapper[4735]: I0215 20:46:41.571664 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" event={"ID":"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40","Type":"ContainerStarted","Data":"bed97d51042ea0f3b0b69a73df11d7b62babde62782ce8c14ede0980ceee725e"} Feb 15 20:46:41 crc kubenswrapper[4735]: I0215 20:46:41.594411 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" podStartSLOduration=2.128890418 podStartE2EDuration="2.594389882s" podCreationTimestamp="2026-02-15 20:46:39 +0000 UTC" firstStartedPulling="2026-02-15 20:46:40.424092132 +0000 UTC m=+1808.290107795" lastFinishedPulling="2026-02-15 20:46:40.889591596 +0000 UTC m=+1808.755607259" observedRunningTime="2026-02-15 20:46:41.587269752 +0000 UTC m=+1809.453285375" watchObservedRunningTime="2026-02-15 20:46:41.594389882 +0000 UTC m=+1809.460405515" Feb 15 20:46:42 crc kubenswrapper[4735]: I0215 20:46:42.898657 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:46:42 crc kubenswrapper[4735]: E0215 20:46:42.899989 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:46:46 crc kubenswrapper[4735]: I0215 20:46:46.626221 4735 generic.go:334] "Generic (PLEG): container finished" podID="6c7b61e8-5a6a-4651-bf3b-3f90ae858b40" containerID="bed97d51042ea0f3b0b69a73df11d7b62babde62782ce8c14ede0980ceee725e" exitCode=0 Feb 15 20:46:46 crc kubenswrapper[4735]: I0215 20:46:46.626436 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" event={"ID":"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40","Type":"ContainerDied","Data":"bed97d51042ea0f3b0b69a73df11d7b62babde62782ce8c14ede0980ceee725e"} Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.051344 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.224166 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-ssh-key-openstack-edpm-ipam\") pod \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.224255 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7srw\" (UniqueName: \"kubernetes.io/projected/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-kube-api-access-k7srw\") pod \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.224293 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-inventory\") pod \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\" (UID: \"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40\") " Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.232649 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-kube-api-access-k7srw" (OuterVolumeSpecName: "kube-api-access-k7srw") pod "6c7b61e8-5a6a-4651-bf3b-3f90ae858b40" (UID: "6c7b61e8-5a6a-4651-bf3b-3f90ae858b40"). InnerVolumeSpecName "kube-api-access-k7srw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.246888 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-inventory" (OuterVolumeSpecName: "inventory") pod "6c7b61e8-5a6a-4651-bf3b-3f90ae858b40" (UID: "6c7b61e8-5a6a-4651-bf3b-3f90ae858b40"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.253839 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6c7b61e8-5a6a-4651-bf3b-3f90ae858b40" (UID: "6c7b61e8-5a6a-4651-bf3b-3f90ae858b40"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.325907 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.325935 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7srw\" (UniqueName: \"kubernetes.io/projected/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-kube-api-access-k7srw\") on node \"crc\" DevicePath \"\"" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.325962 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c7b61e8-5a6a-4651-bf3b-3f90ae858b40-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.644764 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" event={"ID":"6c7b61e8-5a6a-4651-bf3b-3f90ae858b40","Type":"ContainerDied","Data":"3c3f66627611dc30b567b5afd8627c12e54e3a6e28d4e5702c54d0a909b346ad"} Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.644806 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c3f66627611dc30b567b5afd8627c12e54e3a6e28d4e5702c54d0a909b346ad" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.644837 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.743247 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc"] Feb 15 20:46:48 crc kubenswrapper[4735]: E0215 20:46:48.743647 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7b61e8-5a6a-4651-bf3b-3f90ae858b40" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.743669 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7b61e8-5a6a-4651-bf3b-3f90ae858b40" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.743914 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c7b61e8-5a6a-4651-bf3b-3f90ae858b40" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.744603 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.751605 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.751872 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.752983 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.756403 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc"] Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.765543 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.936058 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c7dpc\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.936159 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8kcg\" (UniqueName: \"kubernetes.io/projected/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-kube-api-access-p8kcg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c7dpc\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:48 crc kubenswrapper[4735]: I0215 20:46:48.936188 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c7dpc\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:49 crc kubenswrapper[4735]: I0215 20:46:49.039308 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c7dpc\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:49 crc kubenswrapper[4735]: I0215 20:46:49.039477 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8kcg\" (UniqueName: \"kubernetes.io/projected/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-kube-api-access-p8kcg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c7dpc\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:49 crc kubenswrapper[4735]: I0215 20:46:49.039516 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c7dpc\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:49 crc kubenswrapper[4735]: I0215 20:46:49.046324 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c7dpc\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:49 crc kubenswrapper[4735]: I0215 20:46:49.048744 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c7dpc\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:49 crc kubenswrapper[4735]: I0215 20:46:49.058507 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8kcg\" (UniqueName: \"kubernetes.io/projected/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-kube-api-access-p8kcg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c7dpc\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:49 crc kubenswrapper[4735]: I0215 20:46:49.074515 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:46:49 crc kubenswrapper[4735]: I0215 20:46:49.663064 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc"] Feb 15 20:46:50 crc kubenswrapper[4735]: I0215 20:46:50.663110 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" event={"ID":"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0","Type":"ContainerStarted","Data":"59384bbbbf519a19b237b3b3c89f1f2a95870af8adadb6d3d89b97a8ff067302"} Feb 15 20:46:50 crc kubenswrapper[4735]: I0215 20:46:50.663480 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" event={"ID":"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0","Type":"ContainerStarted","Data":"cb9a66fcc53a8bfb69f2999cefa6bb4472cbab867d3143ee82590e6f9a1c8ca3"} Feb 15 20:46:50 crc kubenswrapper[4735]: I0215 20:46:50.703801 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" podStartSLOduration=2.226631885 podStartE2EDuration="2.703780991s" podCreationTimestamp="2026-02-15 20:46:48 +0000 UTC" firstStartedPulling="2026-02-15 20:46:49.674780157 +0000 UTC m=+1817.540795770" lastFinishedPulling="2026-02-15 20:46:50.151929253 +0000 UTC m=+1818.017944876" observedRunningTime="2026-02-15 20:46:50.687900945 +0000 UTC m=+1818.553916568" watchObservedRunningTime="2026-02-15 20:46:50.703780991 +0000 UTC m=+1818.569796624" Feb 15 20:46:55 crc kubenswrapper[4735]: I0215 20:46:55.887675 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:46:55 crc kubenswrapper[4735]: E0215 20:46:55.888751 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:47:08 crc kubenswrapper[4735]: I0215 20:47:08.887163 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:47:08 crc kubenswrapper[4735]: E0215 20:47:08.887797 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:47:11 crc kubenswrapper[4735]: I0215 20:47:11.045581 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m8x9h"] Feb 15 20:47:11 crc kubenswrapper[4735]: I0215 20:47:11.053164 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m8x9h"] Feb 15 20:47:12 crc kubenswrapper[4735]: I0215 20:47:12.898858 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddc4d167-d274-40c2-9d44-78bb0446ce79" path="/var/lib/kubelet/pods/ddc4d167-d274-40c2-9d44-78bb0446ce79/volumes" Feb 15 20:47:22 crc kubenswrapper[4735]: I0215 20:47:22.892422 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:47:22 crc kubenswrapper[4735]: E0215 20:47:22.893283 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:47:30 crc kubenswrapper[4735]: I0215 20:47:30.030925 4735 generic.go:334] "Generic (PLEG): container finished" podID="38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0" containerID="59384bbbbf519a19b237b3b3c89f1f2a95870af8adadb6d3d89b97a8ff067302" exitCode=0 Feb 15 20:47:30 crc kubenswrapper[4735]: I0215 20:47:30.031069 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" event={"ID":"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0","Type":"ContainerDied","Data":"59384bbbbf519a19b237b3b3c89f1f2a95870af8adadb6d3d89b97a8ff067302"} Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.435181 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.506100 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-inventory\") pod \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.506414 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8kcg\" (UniqueName: \"kubernetes.io/projected/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-kube-api-access-p8kcg\") pod \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.506599 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-ssh-key-openstack-edpm-ipam\") pod \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\" (UID: \"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0\") " Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.514260 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-kube-api-access-p8kcg" (OuterVolumeSpecName: "kube-api-access-p8kcg") pod "38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0" (UID: "38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0"). InnerVolumeSpecName "kube-api-access-p8kcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.537282 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0" (UID: "38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.543881 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-inventory" (OuterVolumeSpecName: "inventory") pod "38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0" (UID: "38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.607997 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.608028 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:47:31 crc kubenswrapper[4735]: I0215 20:47:31.608038 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8kcg\" (UniqueName: \"kubernetes.io/projected/38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0-kube-api-access-p8kcg\") on node \"crc\" DevicePath \"\"" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.054243 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" event={"ID":"38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0","Type":"ContainerDied","Data":"cb9a66fcc53a8bfb69f2999cefa6bb4472cbab867d3143ee82590e6f9a1c8ca3"} Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.054307 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb9a66fcc53a8bfb69f2999cefa6bb4472cbab867d3143ee82590e6f9a1c8ca3" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.054320 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c7dpc" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.163832 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x"] Feb 15 20:47:32 crc kubenswrapper[4735]: E0215 20:47:32.197616 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.197658 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.197927 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.198968 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x"] Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.199080 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.203606 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.204028 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.204295 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.206220 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.220839 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.221020 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2mc5\" (UniqueName: \"kubernetes.io/projected/4a254833-483c-4f57-8407-2dad01349aec-kube-api-access-m2mc5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.221106 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.322560 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.323208 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2mc5\" (UniqueName: \"kubernetes.io/projected/4a254833-483c-4f57-8407-2dad01349aec-kube-api-access-m2mc5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.323331 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.329609 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.330278 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.342849 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2mc5\" (UniqueName: \"kubernetes.io/projected/4a254833-483c-4f57-8407-2dad01349aec-kube-api-access-m2mc5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:32 crc kubenswrapper[4735]: I0215 20:47:32.535465 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:47:33 crc kubenswrapper[4735]: W0215 20:47:33.104434 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a254833_483c_4f57_8407_2dad01349aec.slice/crio-5b744804c5f8870239acddf7916f94c6a76de4a185e0afecf8cf169347994485 WatchSource:0}: Error finding container 5b744804c5f8870239acddf7916f94c6a76de4a185e0afecf8cf169347994485: Status 404 returned error can't find the container with id 5b744804c5f8870239acddf7916f94c6a76de4a185e0afecf8cf169347994485 Feb 15 20:47:33 crc kubenswrapper[4735]: I0215 20:47:33.115931 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x"] Feb 15 20:47:33 crc kubenswrapper[4735]: I0215 20:47:33.552803 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:47:34 crc kubenswrapper[4735]: I0215 20:47:34.087836 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" event={"ID":"4a254833-483c-4f57-8407-2dad01349aec","Type":"ContainerStarted","Data":"fd92fc39561d55520f866eaeb91d070044f4b435d4cdf43bf81d616d5b171998"} Feb 15 20:47:34 crc kubenswrapper[4735]: I0215 20:47:34.088118 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" event={"ID":"4a254833-483c-4f57-8407-2dad01349aec","Type":"ContainerStarted","Data":"5b744804c5f8870239acddf7916f94c6a76de4a185e0afecf8cf169347994485"} Feb 15 20:47:34 crc kubenswrapper[4735]: I0215 20:47:34.108968 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" podStartSLOduration=1.668342263 podStartE2EDuration="2.10892789s" podCreationTimestamp="2026-02-15 20:47:32 +0000 UTC" firstStartedPulling="2026-02-15 20:47:33.109736704 +0000 UTC m=+1860.975752317" lastFinishedPulling="2026-02-15 20:47:33.550322321 +0000 UTC m=+1861.416337944" observedRunningTime="2026-02-15 20:47:34.105903789 +0000 UTC m=+1861.971919452" watchObservedRunningTime="2026-02-15 20:47:34.10892789 +0000 UTC m=+1861.974943533" Feb 15 20:47:34 crc kubenswrapper[4735]: I0215 20:47:34.887057 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:47:34 crc kubenswrapper[4735]: E0215 20:47:34.887624 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:47:35 crc kubenswrapper[4735]: I0215 20:47:35.060799 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-x6l2z"] Feb 15 20:47:35 crc kubenswrapper[4735]: I0215 20:47:35.076109 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-x6l2z"] Feb 15 20:47:36 crc kubenswrapper[4735]: I0215 20:47:36.905285 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58dfdb77-8b78-4db4-9d54-6e29394adbed" path="/var/lib/kubelet/pods/58dfdb77-8b78-4db4-9d54-6e29394adbed/volumes" Feb 15 20:47:38 crc kubenswrapper[4735]: I0215 20:47:38.033540 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gcj9j"] Feb 15 20:47:38 crc kubenswrapper[4735]: I0215 20:47:38.044759 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gcj9j"] Feb 15 20:47:38 crc kubenswrapper[4735]: I0215 20:47:38.467644 4735 scope.go:117] "RemoveContainer" containerID="26b9cc975a1348b164a681f96d9d5e400fbc845f81921d89ce8555b9b5dcba7a" Feb 15 20:47:38 crc kubenswrapper[4735]: I0215 20:47:38.497966 4735 scope.go:117] "RemoveContainer" containerID="a401b2f954dce5c04bfb89229f50e102b6c7f904370f35a0bce6362e666785ee" Feb 15 20:47:38 crc kubenswrapper[4735]: I0215 20:47:38.547636 4735 scope.go:117] "RemoveContainer" containerID="83c1e4a202bd81bfb86878c03dfdc78c9f43f793413d3a965a726ae68500d2a6" Feb 15 20:47:38 crc kubenswrapper[4735]: I0215 20:47:38.578125 4735 scope.go:117] "RemoveContainer" containerID="86d3fafc558c986f2aeaf03b0f5c1adbad067a1996512df8a426bab8ee40a88e" Feb 15 20:47:38 crc kubenswrapper[4735]: I0215 20:47:38.646200 4735 scope.go:117] "RemoveContainer" containerID="8a23895043faca9a454ecf4694ebaee4530963d9005db42b7838e1641d8c4a1c" Feb 15 20:47:38 crc kubenswrapper[4735]: I0215 20:47:38.675001 4735 scope.go:117] "RemoveContainer" containerID="c30cb3b3720eae317ee91f199f741bd74fe72b9894661fd98f2c3bbb66e16e83" Feb 15 20:47:38 crc kubenswrapper[4735]: I0215 20:47:38.898272 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="487d79f2-4901-40f7-8fa4-8b38e39cfb40" path="/var/lib/kubelet/pods/487d79f2-4901-40f7-8fa4-8b38e39cfb40/volumes" Feb 15 20:47:47 crc kubenswrapper[4735]: I0215 20:47:47.887543 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:47:47 crc kubenswrapper[4735]: E0215 20:47:47.888172 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:48:00 crc kubenswrapper[4735]: I0215 20:48:00.887119 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:48:00 crc kubenswrapper[4735]: E0215 20:48:00.887691 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:48:14 crc kubenswrapper[4735]: I0215 20:48:14.887331 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:48:14 crc kubenswrapper[4735]: E0215 20:48:14.888391 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:48:17 crc kubenswrapper[4735]: I0215 20:48:17.613133 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" podUID="468178fa-6b3c-41b4-9bf2-45178ebfb3c2" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.72:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 15 20:48:17 crc kubenswrapper[4735]: I0215 20:48:17.613271 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-webhook-687f57d79b-jvntw" podUID="468178fa-6b3c-41b4-9bf2-45178ebfb3c2" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.72:6080/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 15 20:48:19 crc kubenswrapper[4735]: I0215 20:48:19.068666 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-9plzv"] Feb 15 20:48:19 crc kubenswrapper[4735]: I0215 20:48:19.077751 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-9plzv"] Feb 15 20:48:20 crc kubenswrapper[4735]: I0215 20:48:20.907810 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99051099-c063-4a04-a4a6-7f0e37e5c424" path="/var/lib/kubelet/pods/99051099-c063-4a04-a4a6-7f0e37e5c424/volumes" Feb 15 20:48:25 crc kubenswrapper[4735]: I0215 20:48:25.886745 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:48:26 crc kubenswrapper[4735]: I0215 20:48:26.782418 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"728ffce9a88b0b401f4b8e2c49fe36be9cdd7e5f2f6a5c64dec18e527e842d18"} Feb 15 20:48:29 crc kubenswrapper[4735]: I0215 20:48:29.806576 4735 generic.go:334] "Generic (PLEG): container finished" podID="4a254833-483c-4f57-8407-2dad01349aec" containerID="fd92fc39561d55520f866eaeb91d070044f4b435d4cdf43bf81d616d5b171998" exitCode=0 Feb 15 20:48:29 crc kubenswrapper[4735]: I0215 20:48:29.806697 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" event={"ID":"4a254833-483c-4f57-8407-2dad01349aec","Type":"ContainerDied","Data":"fd92fc39561d55520f866eaeb91d070044f4b435d4cdf43bf81d616d5b171998"} Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.246310 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.338473 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-inventory\") pod \"4a254833-483c-4f57-8407-2dad01349aec\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.338570 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-ssh-key-openstack-edpm-ipam\") pod \"4a254833-483c-4f57-8407-2dad01349aec\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.338694 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2mc5\" (UniqueName: \"kubernetes.io/projected/4a254833-483c-4f57-8407-2dad01349aec-kube-api-access-m2mc5\") pod \"4a254833-483c-4f57-8407-2dad01349aec\" (UID: \"4a254833-483c-4f57-8407-2dad01349aec\") " Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.343732 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a254833-483c-4f57-8407-2dad01349aec-kube-api-access-m2mc5" (OuterVolumeSpecName: "kube-api-access-m2mc5") pod "4a254833-483c-4f57-8407-2dad01349aec" (UID: "4a254833-483c-4f57-8407-2dad01349aec"). InnerVolumeSpecName "kube-api-access-m2mc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.366715 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4a254833-483c-4f57-8407-2dad01349aec" (UID: "4a254833-483c-4f57-8407-2dad01349aec"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.390927 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-inventory" (OuterVolumeSpecName: "inventory") pod "4a254833-483c-4f57-8407-2dad01349aec" (UID: "4a254833-483c-4f57-8407-2dad01349aec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.440684 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2mc5\" (UniqueName: \"kubernetes.io/projected/4a254833-483c-4f57-8407-2dad01349aec-kube-api-access-m2mc5\") on node \"crc\" DevicePath \"\"" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.440714 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.440723 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a254833-483c-4f57-8407-2dad01349aec-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.823285 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" event={"ID":"4a254833-483c-4f57-8407-2dad01349aec","Type":"ContainerDied","Data":"5b744804c5f8870239acddf7916f94c6a76de4a185e0afecf8cf169347994485"} Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.823327 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b744804c5f8870239acddf7916f94c6a76de4a185e0afecf8cf169347994485" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.823335 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.934553 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-frlr4"] Feb 15 20:48:31 crc kubenswrapper[4735]: E0215 20:48:31.934895 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a254833-483c-4f57-8407-2dad01349aec" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.934912 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a254833-483c-4f57-8407-2dad01349aec" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.935108 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a254833-483c-4f57-8407-2dad01349aec" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.935749 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.938455 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.938467 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.938599 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.940368 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:48:31 crc kubenswrapper[4735]: I0215 20:48:31.959681 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-frlr4"] Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.051133 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vbb7\" (UniqueName: \"kubernetes.io/projected/08f45e83-df1b-430c-a2af-92b9370912a5-kube-api-access-7vbb7\") pod \"ssh-known-hosts-edpm-deployment-frlr4\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.051257 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-frlr4\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.051387 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-frlr4\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.153162 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-frlr4\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.153256 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-frlr4\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.153293 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vbb7\" (UniqueName: \"kubernetes.io/projected/08f45e83-df1b-430c-a2af-92b9370912a5-kube-api-access-7vbb7\") pod \"ssh-known-hosts-edpm-deployment-frlr4\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.156438 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-frlr4\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.158353 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-frlr4\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.176504 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vbb7\" (UniqueName: \"kubernetes.io/projected/08f45e83-df1b-430c-a2af-92b9370912a5-kube-api-access-7vbb7\") pod \"ssh-known-hosts-edpm-deployment-frlr4\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.253940 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.764205 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-frlr4"] Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.770837 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 20:48:32 crc kubenswrapper[4735]: I0215 20:48:32.832559 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" event={"ID":"08f45e83-df1b-430c-a2af-92b9370912a5","Type":"ContainerStarted","Data":"295d1259d229bed9d241991d61302b8e522db15167c71e30d1476e257ca7b216"} Feb 15 20:48:33 crc kubenswrapper[4735]: I0215 20:48:33.176994 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:48:33 crc kubenswrapper[4735]: I0215 20:48:33.841492 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" event={"ID":"08f45e83-df1b-430c-a2af-92b9370912a5","Type":"ContainerStarted","Data":"ab321ad61c1c25a984da6433db26c82b8cecb36217601b61bb0aeb47e8aac63c"} Feb 15 20:48:33 crc kubenswrapper[4735]: I0215 20:48:33.904548 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" podStartSLOduration=2.501396772 podStartE2EDuration="2.904524824s" podCreationTimestamp="2026-02-15 20:48:31 +0000 UTC" firstStartedPulling="2026-02-15 20:48:32.770651619 +0000 UTC m=+1920.636667242" lastFinishedPulling="2026-02-15 20:48:33.173779671 +0000 UTC m=+1921.039795294" observedRunningTime="2026-02-15 20:48:33.861894421 +0000 UTC m=+1921.727910064" watchObservedRunningTime="2026-02-15 20:48:33.904524824 +0000 UTC m=+1921.770540447" Feb 15 20:48:38 crc kubenswrapper[4735]: I0215 20:48:38.835295 4735 scope.go:117] "RemoveContainer" containerID="e333e9e509fcf421e4db4b2a9404903962b885dff8678be3b722771fc90fd4ed" Feb 15 20:48:38 crc kubenswrapper[4735]: I0215 20:48:38.880803 4735 scope.go:117] "RemoveContainer" containerID="f28691c3b1cafc1c9b500be9d06d790910e02838248da570947e81cb0b2917fd" Feb 15 20:48:40 crc kubenswrapper[4735]: I0215 20:48:40.906835 4735 generic.go:334] "Generic (PLEG): container finished" podID="08f45e83-df1b-430c-a2af-92b9370912a5" containerID="ab321ad61c1c25a984da6433db26c82b8cecb36217601b61bb0aeb47e8aac63c" exitCode=0 Feb 15 20:48:40 crc kubenswrapper[4735]: I0215 20:48:40.913006 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" event={"ID":"08f45e83-df1b-430c-a2af-92b9370912a5","Type":"ContainerDied","Data":"ab321ad61c1c25a984da6433db26c82b8cecb36217601b61bb0aeb47e8aac63c"} Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.381109 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.465391 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vbb7\" (UniqueName: \"kubernetes.io/projected/08f45e83-df1b-430c-a2af-92b9370912a5-kube-api-access-7vbb7\") pod \"08f45e83-df1b-430c-a2af-92b9370912a5\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.465551 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-ssh-key-openstack-edpm-ipam\") pod \"08f45e83-df1b-430c-a2af-92b9370912a5\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.465583 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-inventory-0\") pod \"08f45e83-df1b-430c-a2af-92b9370912a5\" (UID: \"08f45e83-df1b-430c-a2af-92b9370912a5\") " Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.474762 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08f45e83-df1b-430c-a2af-92b9370912a5-kube-api-access-7vbb7" (OuterVolumeSpecName: "kube-api-access-7vbb7") pod "08f45e83-df1b-430c-a2af-92b9370912a5" (UID: "08f45e83-df1b-430c-a2af-92b9370912a5"). InnerVolumeSpecName "kube-api-access-7vbb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.492619 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "08f45e83-df1b-430c-a2af-92b9370912a5" (UID: "08f45e83-df1b-430c-a2af-92b9370912a5"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.498202 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "08f45e83-df1b-430c-a2af-92b9370912a5" (UID: "08f45e83-df1b-430c-a2af-92b9370912a5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.568391 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vbb7\" (UniqueName: \"kubernetes.io/projected/08f45e83-df1b-430c-a2af-92b9370912a5-kube-api-access-7vbb7\") on node \"crc\" DevicePath \"\"" Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.568438 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.568458 4735 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/08f45e83-df1b-430c-a2af-92b9370912a5-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.932101 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" event={"ID":"08f45e83-df1b-430c-a2af-92b9370912a5","Type":"ContainerDied","Data":"295d1259d229bed9d241991d61302b8e522db15167c71e30d1476e257ca7b216"} Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.932154 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="295d1259d229bed9d241991d61302b8e522db15167c71e30d1476e257ca7b216" Feb 15 20:48:42 crc kubenswrapper[4735]: I0215 20:48:42.932411 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-frlr4" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.022205 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7"] Feb 15 20:48:43 crc kubenswrapper[4735]: E0215 20:48:43.022856 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f45e83-df1b-430c-a2af-92b9370912a5" containerName="ssh-known-hosts-edpm-deployment" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.022967 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f45e83-df1b-430c-a2af-92b9370912a5" containerName="ssh-known-hosts-edpm-deployment" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.023294 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="08f45e83-df1b-430c-a2af-92b9370912a5" containerName="ssh-known-hosts-edpm-deployment" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.024105 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.030129 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.030580 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.031374 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.031644 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.041323 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7"] Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.179521 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9wkf7\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.179586 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9wkf7\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.180150 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vffdb\" (UniqueName: \"kubernetes.io/projected/a7ab7547-fa81-411c-b9dd-fed0f819cd41-kube-api-access-vffdb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9wkf7\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.281496 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9wkf7\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.281736 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9wkf7\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.281927 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vffdb\" (UniqueName: \"kubernetes.io/projected/a7ab7547-fa81-411c-b9dd-fed0f819cd41-kube-api-access-vffdb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9wkf7\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.286129 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9wkf7\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.286129 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9wkf7\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.299997 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vffdb\" (UniqueName: \"kubernetes.io/projected/a7ab7547-fa81-411c-b9dd-fed0f819cd41-kube-api-access-vffdb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9wkf7\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.341963 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.900147 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7"] Feb 15 20:48:43 crc kubenswrapper[4735]: I0215 20:48:43.941461 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" event={"ID":"a7ab7547-fa81-411c-b9dd-fed0f819cd41","Type":"ContainerStarted","Data":"f0f6ae5d7e2ba20a140e506ba7e46bdb8b39e2ca340f086410e86f01dceb57d1"} Feb 15 20:48:44 crc kubenswrapper[4735]: I0215 20:48:44.963002 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" event={"ID":"a7ab7547-fa81-411c-b9dd-fed0f819cd41","Type":"ContainerStarted","Data":"4c3cc4a5b45f068b0c8f4596de6767241743e0099a50adf1b948cf0f7c2c2d30"} Feb 15 20:48:44 crc kubenswrapper[4735]: I0215 20:48:44.980824 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" podStartSLOduration=2.5530691880000003 podStartE2EDuration="2.980806491s" podCreationTimestamp="2026-02-15 20:48:42 +0000 UTC" firstStartedPulling="2026-02-15 20:48:43.914188648 +0000 UTC m=+1931.780204281" lastFinishedPulling="2026-02-15 20:48:44.341925951 +0000 UTC m=+1932.207941584" observedRunningTime="2026-02-15 20:48:44.979476096 +0000 UTC m=+1932.845491719" watchObservedRunningTime="2026-02-15 20:48:44.980806491 +0000 UTC m=+1932.846822114" Feb 15 20:48:54 crc kubenswrapper[4735]: I0215 20:48:54.040556 4735 generic.go:334] "Generic (PLEG): container finished" podID="a7ab7547-fa81-411c-b9dd-fed0f819cd41" containerID="4c3cc4a5b45f068b0c8f4596de6767241743e0099a50adf1b948cf0f7c2c2d30" exitCode=0 Feb 15 20:48:54 crc kubenswrapper[4735]: I0215 20:48:54.040639 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" event={"ID":"a7ab7547-fa81-411c-b9dd-fed0f819cd41","Type":"ContainerDied","Data":"4c3cc4a5b45f068b0c8f4596de6767241743e0099a50adf1b948cf0f7c2c2d30"} Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.443110 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.540270 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-ssh-key-openstack-edpm-ipam\") pod \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.540318 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-inventory\") pod \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.540432 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vffdb\" (UniqueName: \"kubernetes.io/projected/a7ab7547-fa81-411c-b9dd-fed0f819cd41-kube-api-access-vffdb\") pod \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\" (UID: \"a7ab7547-fa81-411c-b9dd-fed0f819cd41\") " Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.545522 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ab7547-fa81-411c-b9dd-fed0f819cd41-kube-api-access-vffdb" (OuterVolumeSpecName: "kube-api-access-vffdb") pod "a7ab7547-fa81-411c-b9dd-fed0f819cd41" (UID: "a7ab7547-fa81-411c-b9dd-fed0f819cd41"). InnerVolumeSpecName "kube-api-access-vffdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.574609 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a7ab7547-fa81-411c-b9dd-fed0f819cd41" (UID: "a7ab7547-fa81-411c-b9dd-fed0f819cd41"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.575373 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-inventory" (OuterVolumeSpecName: "inventory") pod "a7ab7547-fa81-411c-b9dd-fed0f819cd41" (UID: "a7ab7547-fa81-411c-b9dd-fed0f819cd41"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.646408 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.646442 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ab7547-fa81-411c-b9dd-fed0f819cd41-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:48:55 crc kubenswrapper[4735]: I0215 20:48:55.646453 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vffdb\" (UniqueName: \"kubernetes.io/projected/a7ab7547-fa81-411c-b9dd-fed0f819cd41-kube-api-access-vffdb\") on node \"crc\" DevicePath \"\"" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.060771 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" event={"ID":"a7ab7547-fa81-411c-b9dd-fed0f819cd41","Type":"ContainerDied","Data":"f0f6ae5d7e2ba20a140e506ba7e46bdb8b39e2ca340f086410e86f01dceb57d1"} Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.061133 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0f6ae5d7e2ba20a140e506ba7e46bdb8b39e2ca340f086410e86f01dceb57d1" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.060865 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9wkf7" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.142770 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf"] Feb 15 20:48:56 crc kubenswrapper[4735]: E0215 20:48:56.143228 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ab7547-fa81-411c-b9dd-fed0f819cd41" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.143251 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ab7547-fa81-411c-b9dd-fed0f819cd41" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.143448 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ab7547-fa81-411c-b9dd-fed0f819cd41" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.144174 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.146145 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.146299 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.146337 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.146924 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.167309 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf"] Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.262062 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.262151 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.262581 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4dkx\" (UniqueName: \"kubernetes.io/projected/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-kube-api-access-b4dkx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.365441 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.365543 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.365652 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4dkx\" (UniqueName: \"kubernetes.io/projected/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-kube-api-access-b4dkx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.372434 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.381841 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.390602 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4dkx\" (UniqueName: \"kubernetes.io/projected/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-kube-api-access-b4dkx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.469048 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:48:56 crc kubenswrapper[4735]: I0215 20:48:56.992861 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf"] Feb 15 20:48:57 crc kubenswrapper[4735]: I0215 20:48:57.070255 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" event={"ID":"7bbcab5d-3968-4e71-b53f-0c78e86b21cb","Type":"ContainerStarted","Data":"d7b820adb68c7c28638f1c09e35d26bc9d25e773d9b6fa6a1dceeb3e182933fc"} Feb 15 20:48:58 crc kubenswrapper[4735]: I0215 20:48:58.079661 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" event={"ID":"7bbcab5d-3968-4e71-b53f-0c78e86b21cb","Type":"ContainerStarted","Data":"b94ad7f81b9c9bdfb607ab8bae0db601f0ca8e9faca9d407634f6aa383250290"} Feb 15 20:49:07 crc kubenswrapper[4735]: I0215 20:49:07.182919 4735 generic.go:334] "Generic (PLEG): container finished" podID="7bbcab5d-3968-4e71-b53f-0c78e86b21cb" containerID="b94ad7f81b9c9bdfb607ab8bae0db601f0ca8e9faca9d407634f6aa383250290" exitCode=0 Feb 15 20:49:07 crc kubenswrapper[4735]: I0215 20:49:07.183011 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" event={"ID":"7bbcab5d-3968-4e71-b53f-0c78e86b21cb","Type":"ContainerDied","Data":"b94ad7f81b9c9bdfb607ab8bae0db601f0ca8e9faca9d407634f6aa383250290"} Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.608105 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.714746 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-ssh-key-openstack-edpm-ipam\") pod \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.714805 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-inventory\") pod \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.714830 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4dkx\" (UniqueName: \"kubernetes.io/projected/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-kube-api-access-b4dkx\") pod \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\" (UID: \"7bbcab5d-3968-4e71-b53f-0c78e86b21cb\") " Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.724645 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-kube-api-access-b4dkx" (OuterVolumeSpecName: "kube-api-access-b4dkx") pod "7bbcab5d-3968-4e71-b53f-0c78e86b21cb" (UID: "7bbcab5d-3968-4e71-b53f-0c78e86b21cb"). InnerVolumeSpecName "kube-api-access-b4dkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.746268 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-inventory" (OuterVolumeSpecName: "inventory") pod "7bbcab5d-3968-4e71-b53f-0c78e86b21cb" (UID: "7bbcab5d-3968-4e71-b53f-0c78e86b21cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.758387 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7bbcab5d-3968-4e71-b53f-0c78e86b21cb" (UID: "7bbcab5d-3968-4e71-b53f-0c78e86b21cb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.816493 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.816524 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:08 crc kubenswrapper[4735]: I0215 20:49:08.816533 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4dkx\" (UniqueName: \"kubernetes.io/projected/7bbcab5d-3968-4e71-b53f-0c78e86b21cb-kube-api-access-b4dkx\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.207063 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" event={"ID":"7bbcab5d-3968-4e71-b53f-0c78e86b21cb","Type":"ContainerDied","Data":"d7b820adb68c7c28638f1c09e35d26bc9d25e773d9b6fa6a1dceeb3e182933fc"} Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.207803 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7b820adb68c7c28638f1c09e35d26bc9d25e773d9b6fa6a1dceeb3e182933fc" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.207229 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.336664 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv"] Feb 15 20:49:09 crc kubenswrapper[4735]: E0215 20:49:09.338872 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bbcab5d-3968-4e71-b53f-0c78e86b21cb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.338898 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bbcab5d-3968-4e71-b53f-0c78e86b21cb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.339318 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bbcab5d-3968-4e71-b53f-0c78e86b21cb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.340251 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.352494 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.352796 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.353036 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.353205 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.353344 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.353472 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.353697 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.353993 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.397040 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv"] Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.431673 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.431726 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.431769 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.431791 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.431967 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n8nc\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-kube-api-access-5n8nc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.432048 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.432183 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.432250 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.432284 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.432336 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.432428 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.432491 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.432512 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.432603 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.533931 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534034 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534085 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534110 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534171 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n8nc\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-kube-api-access-5n8nc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534212 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534267 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534310 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534341 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534382 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534435 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534474 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534498 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.534529 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.539480 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.539788 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.540171 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.540319 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.540611 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.541303 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.541784 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.542796 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.543868 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.543890 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.544604 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.544650 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.545579 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.552776 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n8nc\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-kube-api-access-5n8nc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:09 crc kubenswrapper[4735]: I0215 20:49:09.665984 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:10 crc kubenswrapper[4735]: I0215 20:49:10.164694 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv"] Feb 15 20:49:10 crc kubenswrapper[4735]: I0215 20:49:10.217976 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" event={"ID":"0ea9cfe8-7371-4d83-8e57-755d86809d46","Type":"ContainerStarted","Data":"657f7e27733503fcdf413977e270df8142af395ecd8e65c88d2263a31647ad6b"} Feb 15 20:49:11 crc kubenswrapper[4735]: I0215 20:49:11.227424 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" event={"ID":"0ea9cfe8-7371-4d83-8e57-755d86809d46","Type":"ContainerStarted","Data":"67561f49bfd6e07897b9f2f657ab48eb53462926c2ee871ab25955a650e325b8"} Feb 15 20:49:11 crc kubenswrapper[4735]: I0215 20:49:11.256533 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" podStartSLOduration=1.79191111 podStartE2EDuration="2.256509961s" podCreationTimestamp="2026-02-15 20:49:09 +0000 UTC" firstStartedPulling="2026-02-15 20:49:10.174887824 +0000 UTC m=+1958.040903447" lastFinishedPulling="2026-02-15 20:49:10.639486655 +0000 UTC m=+1958.505502298" observedRunningTime="2026-02-15 20:49:11.250215133 +0000 UTC m=+1959.116230766" watchObservedRunningTime="2026-02-15 20:49:11.256509961 +0000 UTC m=+1959.122525594" Feb 15 20:49:48 crc kubenswrapper[4735]: I0215 20:49:48.617884 4735 generic.go:334] "Generic (PLEG): container finished" podID="0ea9cfe8-7371-4d83-8e57-755d86809d46" containerID="67561f49bfd6e07897b9f2f657ab48eb53462926c2ee871ab25955a650e325b8" exitCode=0 Feb 15 20:49:48 crc kubenswrapper[4735]: I0215 20:49:48.618000 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" event={"ID":"0ea9cfe8-7371-4d83-8e57-755d86809d46","Type":"ContainerDied","Data":"67561f49bfd6e07897b9f2f657ab48eb53462926c2ee871ab25955a650e325b8"} Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.059123 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.198793 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-inventory\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.198908 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-ovn-default-certs-0\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199036 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-neutron-metadata-combined-ca-bundle\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199071 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-telemetry-combined-ca-bundle\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199110 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-repo-setup-combined-ca-bundle\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199140 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ssh-key-openstack-edpm-ipam\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199170 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-nova-combined-ca-bundle\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199196 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-libvirt-combined-ca-bundle\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199234 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199264 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199341 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199430 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-bootstrap-combined-ca-bundle\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199475 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n8nc\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-kube-api-access-5n8nc\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.199526 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ovn-combined-ca-bundle\") pod \"0ea9cfe8-7371-4d83-8e57-755d86809d46\" (UID: \"0ea9cfe8-7371-4d83-8e57-755d86809d46\") " Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.206448 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.206478 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.206502 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.207483 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.207572 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.209061 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.210386 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.210640 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.212263 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-kube-api-access-5n8nc" (OuterVolumeSpecName: "kube-api-access-5n8nc") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "kube-api-access-5n8nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.212649 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.212720 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.213472 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.243076 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-inventory" (OuterVolumeSpecName: "inventory") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.246584 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0ea9cfe8-7371-4d83-8e57-755d86809d46" (UID: "0ea9cfe8-7371-4d83-8e57-755d86809d46"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301538 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301592 4735 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301614 4735 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301633 4735 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301652 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301670 4735 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301687 4735 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301705 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301727 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301747 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301765 4735 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301783 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n8nc\" (UniqueName: \"kubernetes.io/projected/0ea9cfe8-7371-4d83-8e57-755d86809d46-kube-api-access-5n8nc\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301802 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.301843 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea9cfe8-7371-4d83-8e57-755d86809d46-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.636358 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" event={"ID":"0ea9cfe8-7371-4d83-8e57-755d86809d46","Type":"ContainerDied","Data":"657f7e27733503fcdf413977e270df8142af395ecd8e65c88d2263a31647ad6b"} Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.636406 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="657f7e27733503fcdf413977e270df8142af395ecd8e65c88d2263a31647ad6b" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.636446 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.780865 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw"] Feb 15 20:49:50 crc kubenswrapper[4735]: E0215 20:49:50.781228 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea9cfe8-7371-4d83-8e57-755d86809d46" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.781246 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea9cfe8-7371-4d83-8e57-755d86809d46" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.781428 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ea9cfe8-7371-4d83-8e57-755d86809d46" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.782060 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.784852 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.785149 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.785474 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.785780 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.798147 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.805201 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw"] Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.913864 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25gv6\" (UniqueName: \"kubernetes.io/projected/5779edb8-01f9-4f91-81fe-9af70711a9bc-kube-api-access-25gv6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.913938 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.913977 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.914048 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:50 crc kubenswrapper[4735]: I0215 20:49:50.914073 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.016011 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.016069 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.016185 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25gv6\" (UniqueName: \"kubernetes.io/projected/5779edb8-01f9-4f91-81fe-9af70711a9bc-kube-api-access-25gv6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.016225 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.016249 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.017146 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.021469 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.021683 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.032417 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.035846 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25gv6\" (UniqueName: \"kubernetes.io/projected/5779edb8-01f9-4f91-81fe-9af70711a9bc-kube-api-access-25gv6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qc8mw\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.104407 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:49:51 crc kubenswrapper[4735]: I0215 20:49:51.668137 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw"] Feb 15 20:49:52 crc kubenswrapper[4735]: I0215 20:49:52.655179 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" event={"ID":"5779edb8-01f9-4f91-81fe-9af70711a9bc","Type":"ContainerStarted","Data":"5d73875173b248055511565c82b2cd9b9a68ab55616d39b0320c274e55f7385c"} Feb 15 20:49:52 crc kubenswrapper[4735]: I0215 20:49:52.656100 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" event={"ID":"5779edb8-01f9-4f91-81fe-9af70711a9bc","Type":"ContainerStarted","Data":"fb768ecd1eeae52d0bd2c4388e172807fe08f3cb14f8ee1a4c49a208fe24bf9a"} Feb 15 20:49:52 crc kubenswrapper[4735]: I0215 20:49:52.676959 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" podStartSLOduration=2.290566024 podStartE2EDuration="2.676929906s" podCreationTimestamp="2026-02-15 20:49:50 +0000 UTC" firstStartedPulling="2026-02-15 20:49:51.688843857 +0000 UTC m=+1999.554859480" lastFinishedPulling="2026-02-15 20:49:52.075207739 +0000 UTC m=+1999.941223362" observedRunningTime="2026-02-15 20:49:52.670040714 +0000 UTC m=+2000.536056337" watchObservedRunningTime="2026-02-15 20:49:52.676929906 +0000 UTC m=+2000.542945529" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.087516 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dsdc8"] Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.090015 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.106242 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dsdc8"] Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.183910 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-utilities\") pod \"redhat-operators-dsdc8\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.184276 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-catalog-content\") pod \"redhat-operators-dsdc8\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.184310 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dc57\" (UniqueName: \"kubernetes.io/projected/2996efaf-7fbd-4881-be23-498ed59db021-kube-api-access-4dc57\") pod \"redhat-operators-dsdc8\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.285576 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-utilities\") pod \"redhat-operators-dsdc8\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.285658 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-catalog-content\") pod \"redhat-operators-dsdc8\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.285689 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dc57\" (UniqueName: \"kubernetes.io/projected/2996efaf-7fbd-4881-be23-498ed59db021-kube-api-access-4dc57\") pod \"redhat-operators-dsdc8\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.286328 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-utilities\") pod \"redhat-operators-dsdc8\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.286483 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-catalog-content\") pod \"redhat-operators-dsdc8\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.305537 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dc57\" (UniqueName: \"kubernetes.io/projected/2996efaf-7fbd-4881-be23-498ed59db021-kube-api-access-4dc57\") pod \"redhat-operators-dsdc8\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.411914 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.837253 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dsdc8"] Feb 15 20:50:28 crc kubenswrapper[4735]: I0215 20:50:28.965668 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dsdc8" event={"ID":"2996efaf-7fbd-4881-be23-498ed59db021","Type":"ContainerStarted","Data":"bda1f214bea92c9296100370f2ad24ed3bb3a2fe43d409365b86163ead76444b"} Feb 15 20:50:29 crc kubenswrapper[4735]: I0215 20:50:29.988466 4735 generic.go:334] "Generic (PLEG): container finished" podID="2996efaf-7fbd-4881-be23-498ed59db021" containerID="fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878" exitCode=0 Feb 15 20:50:29 crc kubenswrapper[4735]: I0215 20:50:29.988625 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dsdc8" event={"ID":"2996efaf-7fbd-4881-be23-498ed59db021","Type":"ContainerDied","Data":"fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878"} Feb 15 20:50:30 crc kubenswrapper[4735]: I0215 20:50:30.998677 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dsdc8" event={"ID":"2996efaf-7fbd-4881-be23-498ed59db021","Type":"ContainerStarted","Data":"0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386"} Feb 15 20:50:36 crc kubenswrapper[4735]: I0215 20:50:36.044011 4735 generic.go:334] "Generic (PLEG): container finished" podID="2996efaf-7fbd-4881-be23-498ed59db021" containerID="0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386" exitCode=0 Feb 15 20:50:36 crc kubenswrapper[4735]: I0215 20:50:36.044116 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dsdc8" event={"ID":"2996efaf-7fbd-4881-be23-498ed59db021","Type":"ContainerDied","Data":"0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386"} Feb 15 20:50:37 crc kubenswrapper[4735]: I0215 20:50:37.054739 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dsdc8" event={"ID":"2996efaf-7fbd-4881-be23-498ed59db021","Type":"ContainerStarted","Data":"ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6"} Feb 15 20:50:37 crc kubenswrapper[4735]: I0215 20:50:37.082532 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dsdc8" podStartSLOduration=2.677634202 podStartE2EDuration="9.082512289s" podCreationTimestamp="2026-02-15 20:50:28 +0000 UTC" firstStartedPulling="2026-02-15 20:50:29.992527147 +0000 UTC m=+2037.858542770" lastFinishedPulling="2026-02-15 20:50:36.397405224 +0000 UTC m=+2044.263420857" observedRunningTime="2026-02-15 20:50:37.072827881 +0000 UTC m=+2044.938843524" watchObservedRunningTime="2026-02-15 20:50:37.082512289 +0000 UTC m=+2044.948527912" Feb 15 20:50:38 crc kubenswrapper[4735]: I0215 20:50:38.416978 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:38 crc kubenswrapper[4735]: I0215 20:50:38.417033 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:50:39 crc kubenswrapper[4735]: I0215 20:50:39.472726 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dsdc8" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="registry-server" probeResult="failure" output=< Feb 15 20:50:39 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:50:39 crc kubenswrapper[4735]: > Feb 15 20:50:49 crc kubenswrapper[4735]: I0215 20:50:49.457177 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dsdc8" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="registry-server" probeResult="failure" output=< Feb 15 20:50:49 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:50:49 crc kubenswrapper[4735]: > Feb 15 20:50:49 crc kubenswrapper[4735]: I0215 20:50:49.679475 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:50:49 crc kubenswrapper[4735]: I0215 20:50:49.679523 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:50:59 crc kubenswrapper[4735]: I0215 20:50:59.479656 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dsdc8" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="registry-server" probeResult="failure" output=< Feb 15 20:50:59 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:50:59 crc kubenswrapper[4735]: > Feb 15 20:51:02 crc kubenswrapper[4735]: I0215 20:51:02.257593 4735 generic.go:334] "Generic (PLEG): container finished" podID="5779edb8-01f9-4f91-81fe-9af70711a9bc" containerID="5d73875173b248055511565c82b2cd9b9a68ab55616d39b0320c274e55f7385c" exitCode=0 Feb 15 20:51:02 crc kubenswrapper[4735]: I0215 20:51:02.258131 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" event={"ID":"5779edb8-01f9-4f91-81fe-9af70711a9bc","Type":"ContainerDied","Data":"5d73875173b248055511565c82b2cd9b9a68ab55616d39b0320c274e55f7385c"} Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.650209 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.761543 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25gv6\" (UniqueName: \"kubernetes.io/projected/5779edb8-01f9-4f91-81fe-9af70711a9bc-kube-api-access-25gv6\") pod \"5779edb8-01f9-4f91-81fe-9af70711a9bc\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.761624 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ssh-key-openstack-edpm-ipam\") pod \"5779edb8-01f9-4f91-81fe-9af70711a9bc\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.761683 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovn-combined-ca-bundle\") pod \"5779edb8-01f9-4f91-81fe-9af70711a9bc\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.761776 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovncontroller-config-0\") pod \"5779edb8-01f9-4f91-81fe-9af70711a9bc\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.761989 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-inventory\") pod \"5779edb8-01f9-4f91-81fe-9af70711a9bc\" (UID: \"5779edb8-01f9-4f91-81fe-9af70711a9bc\") " Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.768279 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5779edb8-01f9-4f91-81fe-9af70711a9bc" (UID: "5779edb8-01f9-4f91-81fe-9af70711a9bc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.768406 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5779edb8-01f9-4f91-81fe-9af70711a9bc-kube-api-access-25gv6" (OuterVolumeSpecName: "kube-api-access-25gv6") pod "5779edb8-01f9-4f91-81fe-9af70711a9bc" (UID: "5779edb8-01f9-4f91-81fe-9af70711a9bc"). InnerVolumeSpecName "kube-api-access-25gv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.785974 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "5779edb8-01f9-4f91-81fe-9af70711a9bc" (UID: "5779edb8-01f9-4f91-81fe-9af70711a9bc"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.786679 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5779edb8-01f9-4f91-81fe-9af70711a9bc" (UID: "5779edb8-01f9-4f91-81fe-9af70711a9bc"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.792647 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-inventory" (OuterVolumeSpecName: "inventory") pod "5779edb8-01f9-4f91-81fe-9af70711a9bc" (UID: "5779edb8-01f9-4f91-81fe-9af70711a9bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.864467 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.864694 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25gv6\" (UniqueName: \"kubernetes.io/projected/5779edb8-01f9-4f91-81fe-9af70711a9bc-kube-api-access-25gv6\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.864703 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.864712 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:03 crc kubenswrapper[4735]: I0215 20:51:03.864721 4735 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5779edb8-01f9-4f91-81fe-9af70711a9bc-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.286328 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" event={"ID":"5779edb8-01f9-4f91-81fe-9af70711a9bc","Type":"ContainerDied","Data":"fb768ecd1eeae52d0bd2c4388e172807fe08f3cb14f8ee1a4c49a208fe24bf9a"} Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.286386 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb768ecd1eeae52d0bd2c4388e172807fe08f3cb14f8ee1a4c49a208fe24bf9a" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.286406 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qc8mw" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.455876 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs"] Feb 15 20:51:04 crc kubenswrapper[4735]: E0215 20:51:04.456409 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5779edb8-01f9-4f91-81fe-9af70711a9bc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.456436 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5779edb8-01f9-4f91-81fe-9af70711a9bc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.456654 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5779edb8-01f9-4f91-81fe-9af70711a9bc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.457576 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.459789 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.460308 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.460659 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.461771 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.461875 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.462177 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.469456 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs"] Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.577433 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.577532 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.577566 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.577650 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6vlb\" (UniqueName: \"kubernetes.io/projected/82354908-424c-4069-abd7-f6b5ededdf13-kube-api-access-c6vlb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.577716 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.577757 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.685822 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.685886 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.685965 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6vlb\" (UniqueName: \"kubernetes.io/projected/82354908-424c-4069-abd7-f6b5ededdf13-kube-api-access-c6vlb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.686026 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.686061 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.686125 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.697526 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.699003 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.699278 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.711114 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.711573 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.718994 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6vlb\" (UniqueName: \"kubernetes.io/projected/82354908-424c-4069-abd7-f6b5ededdf13-kube-api-access-c6vlb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:04 crc kubenswrapper[4735]: I0215 20:51:04.789195 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:05 crc kubenswrapper[4735]: I0215 20:51:05.359786 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs"] Feb 15 20:51:06 crc kubenswrapper[4735]: I0215 20:51:06.307270 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" event={"ID":"82354908-424c-4069-abd7-f6b5ededdf13","Type":"ContainerStarted","Data":"6aa2d4599134bcb3841ef5ff7d3b8f9fd03143e984fd0a3df934634bc9359004"} Feb 15 20:51:06 crc kubenswrapper[4735]: I0215 20:51:06.307615 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" event={"ID":"82354908-424c-4069-abd7-f6b5ededdf13","Type":"ContainerStarted","Data":"b61d25841c02e87c0610dcab2efaf3e6e9d8783a8f5324ac2a52bfd513d9c098"} Feb 15 20:51:06 crc kubenswrapper[4735]: I0215 20:51:06.327898 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" podStartSLOduration=1.92630704 podStartE2EDuration="2.327878496s" podCreationTimestamp="2026-02-15 20:51:04 +0000 UTC" firstStartedPulling="2026-02-15 20:51:05.362291705 +0000 UTC m=+2073.228307328" lastFinishedPulling="2026-02-15 20:51:05.763863151 +0000 UTC m=+2073.629878784" observedRunningTime="2026-02-15 20:51:06.327315021 +0000 UTC m=+2074.193330644" watchObservedRunningTime="2026-02-15 20:51:06.327878496 +0000 UTC m=+2074.193894129" Feb 15 20:51:08 crc kubenswrapper[4735]: I0215 20:51:08.494580 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:51:08 crc kubenswrapper[4735]: I0215 20:51:08.549025 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:51:08 crc kubenswrapper[4735]: I0215 20:51:08.750409 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dsdc8"] Feb 15 20:51:10 crc kubenswrapper[4735]: I0215 20:51:10.341414 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dsdc8" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="registry-server" containerID="cri-o://ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6" gracePeriod=2 Feb 15 20:51:10 crc kubenswrapper[4735]: I0215 20:51:10.867584 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:51:10 crc kubenswrapper[4735]: I0215 20:51:10.902463 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-utilities\") pod \"2996efaf-7fbd-4881-be23-498ed59db021\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " Feb 15 20:51:10 crc kubenswrapper[4735]: I0215 20:51:10.902512 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-catalog-content\") pod \"2996efaf-7fbd-4881-be23-498ed59db021\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " Feb 15 20:51:10 crc kubenswrapper[4735]: I0215 20:51:10.902597 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dc57\" (UniqueName: \"kubernetes.io/projected/2996efaf-7fbd-4881-be23-498ed59db021-kube-api-access-4dc57\") pod \"2996efaf-7fbd-4881-be23-498ed59db021\" (UID: \"2996efaf-7fbd-4881-be23-498ed59db021\") " Feb 15 20:51:10 crc kubenswrapper[4735]: I0215 20:51:10.903400 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-utilities" (OuterVolumeSpecName: "utilities") pod "2996efaf-7fbd-4881-be23-498ed59db021" (UID: "2996efaf-7fbd-4881-be23-498ed59db021"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:51:10 crc kubenswrapper[4735]: I0215 20:51:10.913631 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2996efaf-7fbd-4881-be23-498ed59db021-kube-api-access-4dc57" (OuterVolumeSpecName: "kube-api-access-4dc57") pod "2996efaf-7fbd-4881-be23-498ed59db021" (UID: "2996efaf-7fbd-4881-be23-498ed59db021"). InnerVolumeSpecName "kube-api-access-4dc57". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.004820 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.004893 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dc57\" (UniqueName: \"kubernetes.io/projected/2996efaf-7fbd-4881-be23-498ed59db021-kube-api-access-4dc57\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.029669 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2996efaf-7fbd-4881-be23-498ed59db021" (UID: "2996efaf-7fbd-4881-be23-498ed59db021"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.107105 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2996efaf-7fbd-4881-be23-498ed59db021-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.355009 4735 generic.go:334] "Generic (PLEG): container finished" podID="2996efaf-7fbd-4881-be23-498ed59db021" containerID="ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6" exitCode=0 Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.355380 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dsdc8" event={"ID":"2996efaf-7fbd-4881-be23-498ed59db021","Type":"ContainerDied","Data":"ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6"} Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.355414 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dsdc8" event={"ID":"2996efaf-7fbd-4881-be23-498ed59db021","Type":"ContainerDied","Data":"bda1f214bea92c9296100370f2ad24ed3bb3a2fe43d409365b86163ead76444b"} Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.355439 4735 scope.go:117] "RemoveContainer" containerID="ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.355594 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dsdc8" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.393937 4735 scope.go:117] "RemoveContainer" containerID="0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.400920 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dsdc8"] Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.417371 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dsdc8"] Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.418326 4735 scope.go:117] "RemoveContainer" containerID="fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.462918 4735 scope.go:117] "RemoveContainer" containerID="ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6" Feb 15 20:51:11 crc kubenswrapper[4735]: E0215 20:51:11.463508 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6\": container with ID starting with ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6 not found: ID does not exist" containerID="ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.463564 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6"} err="failed to get container status \"ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6\": rpc error: code = NotFound desc = could not find container \"ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6\": container with ID starting with ffd5a7f3ae11e4dab4df7c262ac3aceb0a1aca9cc7a54b420776b8ae36289dc6 not found: ID does not exist" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.463590 4735 scope.go:117] "RemoveContainer" containerID="0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386" Feb 15 20:51:11 crc kubenswrapper[4735]: E0215 20:51:11.464534 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386\": container with ID starting with 0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386 not found: ID does not exist" containerID="0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.464581 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386"} err="failed to get container status \"0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386\": rpc error: code = NotFound desc = could not find container \"0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386\": container with ID starting with 0581702826c8bf9ce559f02333f6dd046fe87fa934273aa8c3fdbbff28687386 not found: ID does not exist" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.464597 4735 scope.go:117] "RemoveContainer" containerID="fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878" Feb 15 20:51:11 crc kubenswrapper[4735]: E0215 20:51:11.464933 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878\": container with ID starting with fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878 not found: ID does not exist" containerID="fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878" Feb 15 20:51:11 crc kubenswrapper[4735]: I0215 20:51:11.464997 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878"} err="failed to get container status \"fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878\": rpc error: code = NotFound desc = could not find container \"fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878\": container with ID starting with fe2672a9fce6b82bd69b164c685f9d69ac441d5b9100eac7c68212945cdc8878 not found: ID does not exist" Feb 15 20:51:12 crc kubenswrapper[4735]: I0215 20:51:12.897265 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2996efaf-7fbd-4881-be23-498ed59db021" path="/var/lib/kubelet/pods/2996efaf-7fbd-4881-be23-498ed59db021/volumes" Feb 15 20:51:19 crc kubenswrapper[4735]: I0215 20:51:19.680071 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:51:19 crc kubenswrapper[4735]: I0215 20:51:19.680854 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:51:49 crc kubenswrapper[4735]: I0215 20:51:49.680060 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:51:49 crc kubenswrapper[4735]: I0215 20:51:49.680565 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:51:49 crc kubenswrapper[4735]: I0215 20:51:49.680615 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:51:49 crc kubenswrapper[4735]: I0215 20:51:49.681464 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"728ffce9a88b0b401f4b8e2c49fe36be9cdd7e5f2f6a5c64dec18e527e842d18"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:51:49 crc kubenswrapper[4735]: I0215 20:51:49.681531 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://728ffce9a88b0b401f4b8e2c49fe36be9cdd7e5f2f6a5c64dec18e527e842d18" gracePeriod=600 Feb 15 20:51:50 crc kubenswrapper[4735]: I0215 20:51:50.714057 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="728ffce9a88b0b401f4b8e2c49fe36be9cdd7e5f2f6a5c64dec18e527e842d18" exitCode=0 Feb 15 20:51:50 crc kubenswrapper[4735]: I0215 20:51:50.714575 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"728ffce9a88b0b401f4b8e2c49fe36be9cdd7e5f2f6a5c64dec18e527e842d18"} Feb 15 20:51:50 crc kubenswrapper[4735]: I0215 20:51:50.714605 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639"} Feb 15 20:51:50 crc kubenswrapper[4735]: I0215 20:51:50.714624 4735 scope.go:117] "RemoveContainer" containerID="dd3cc128c87c142101dc0edae357002ed8526beca5a5959921959fb53364e22e" Feb 15 20:51:55 crc kubenswrapper[4735]: E0215 20:51:55.031485 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82354908_424c_4069_abd7_f6b5ededdf13.slice/crio-6aa2d4599134bcb3841ef5ff7d3b8f9fd03143e984fd0a3df934634bc9359004.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82354908_424c_4069_abd7_f6b5ededdf13.slice/crio-conmon-6aa2d4599134bcb3841ef5ff7d3b8f9fd03143e984fd0a3df934634bc9359004.scope\": RecentStats: unable to find data in memory cache]" Feb 15 20:51:55 crc kubenswrapper[4735]: I0215 20:51:55.758822 4735 generic.go:334] "Generic (PLEG): container finished" podID="82354908-424c-4069-abd7-f6b5ededdf13" containerID="6aa2d4599134bcb3841ef5ff7d3b8f9fd03143e984fd0a3df934634bc9359004" exitCode=0 Feb 15 20:51:55 crc kubenswrapper[4735]: I0215 20:51:55.759047 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" event={"ID":"82354908-424c-4069-abd7-f6b5ededdf13","Type":"ContainerDied","Data":"6aa2d4599134bcb3841ef5ff7d3b8f9fd03143e984fd0a3df934634bc9359004"} Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.259257 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.390526 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-inventory\") pod \"82354908-424c-4069-abd7-f6b5ededdf13\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.390697 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-ovn-metadata-agent-neutron-config-0\") pod \"82354908-424c-4069-abd7-f6b5ededdf13\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.390788 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6vlb\" (UniqueName: \"kubernetes.io/projected/82354908-424c-4069-abd7-f6b5ededdf13-kube-api-access-c6vlb\") pod \"82354908-424c-4069-abd7-f6b5ededdf13\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.390819 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-nova-metadata-neutron-config-0\") pod \"82354908-424c-4069-abd7-f6b5ededdf13\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.390860 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-ssh-key-openstack-edpm-ipam\") pod \"82354908-424c-4069-abd7-f6b5ededdf13\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.390913 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-metadata-combined-ca-bundle\") pod \"82354908-424c-4069-abd7-f6b5ededdf13\" (UID: \"82354908-424c-4069-abd7-f6b5ededdf13\") " Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.398141 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "82354908-424c-4069-abd7-f6b5ededdf13" (UID: "82354908-424c-4069-abd7-f6b5ededdf13"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.398973 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82354908-424c-4069-abd7-f6b5ededdf13-kube-api-access-c6vlb" (OuterVolumeSpecName: "kube-api-access-c6vlb") pod "82354908-424c-4069-abd7-f6b5ededdf13" (UID: "82354908-424c-4069-abd7-f6b5ededdf13"). InnerVolumeSpecName "kube-api-access-c6vlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.419465 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "82354908-424c-4069-abd7-f6b5ededdf13" (UID: "82354908-424c-4069-abd7-f6b5ededdf13"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.423187 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "82354908-424c-4069-abd7-f6b5ededdf13" (UID: "82354908-424c-4069-abd7-f6b5ededdf13"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.431994 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "82354908-424c-4069-abd7-f6b5ededdf13" (UID: "82354908-424c-4069-abd7-f6b5ededdf13"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.435855 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-inventory" (OuterVolumeSpecName: "inventory") pod "82354908-424c-4069-abd7-f6b5ededdf13" (UID: "82354908-424c-4069-abd7-f6b5ededdf13"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.493677 4735 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.493742 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6vlb\" (UniqueName: \"kubernetes.io/projected/82354908-424c-4069-abd7-f6b5ededdf13-kube-api-access-c6vlb\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.493766 4735 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.493786 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.493806 4735 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.493827 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82354908-424c-4069-abd7-f6b5ededdf13-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.782363 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" event={"ID":"82354908-424c-4069-abd7-f6b5ededdf13","Type":"ContainerDied","Data":"b61d25841c02e87c0610dcab2efaf3e6e9d8783a8f5324ac2a52bfd513d9c098"} Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.782647 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b61d25841c02e87c0610dcab2efaf3e6e9d8783a8f5324ac2a52bfd513d9c098" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.782461 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.952484 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g"] Feb 15 20:51:57 crc kubenswrapper[4735]: E0215 20:51:57.952842 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82354908-424c-4069-abd7-f6b5ededdf13" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.952859 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="82354908-424c-4069-abd7-f6b5ededdf13" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 15 20:51:57 crc kubenswrapper[4735]: E0215 20:51:57.952871 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="registry-server" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.952879 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="registry-server" Feb 15 20:51:57 crc kubenswrapper[4735]: E0215 20:51:57.952901 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="extract-content" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.952906 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="extract-content" Feb 15 20:51:57 crc kubenswrapper[4735]: E0215 20:51:57.952925 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="extract-utilities" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.952932 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="extract-utilities" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.953119 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2996efaf-7fbd-4881-be23-498ed59db021" containerName="registry-server" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.953134 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="82354908-424c-4069-abd7-f6b5ededdf13" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.953694 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.956093 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.957051 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.960284 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.960663 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.970417 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:51:57 crc kubenswrapper[4735]: I0215 20:51:57.970993 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g"] Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.106216 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.106542 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.106612 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.106696 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.106746 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5wjl\" (UniqueName: \"kubernetes.io/projected/3e63486b-442e-477f-8692-cb3da5e648cd-kube-api-access-d5wjl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.208709 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.208773 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.208798 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.209529 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5wjl\" (UniqueName: \"kubernetes.io/projected/3e63486b-442e-477f-8692-cb3da5e648cd-kube-api-access-d5wjl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.209645 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.213095 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.214191 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.215794 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.218411 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.228670 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5wjl\" (UniqueName: \"kubernetes.io/projected/3e63486b-442e-477f-8692-cb3da5e648cd-kube-api-access-d5wjl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.308324 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:51:58 crc kubenswrapper[4735]: I0215 20:51:58.849442 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g"] Feb 15 20:51:59 crc kubenswrapper[4735]: I0215 20:51:59.813729 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" event={"ID":"3e63486b-442e-477f-8692-cb3da5e648cd","Type":"ContainerStarted","Data":"ce0ecdb4914f3d103027f9eaacc71fe6b9173e1d899efb53ff0134ac79a6e70d"} Feb 15 20:51:59 crc kubenswrapper[4735]: I0215 20:51:59.814285 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" event={"ID":"3e63486b-442e-477f-8692-cb3da5e648cd","Type":"ContainerStarted","Data":"323209fbcba95da9c5a2645693181656b5fd2011a6db912fd0290164a7e5b312"} Feb 15 20:51:59 crc kubenswrapper[4735]: I0215 20:51:59.834387 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" podStartSLOduration=2.356428719 podStartE2EDuration="2.834371435s" podCreationTimestamp="2026-02-15 20:51:57 +0000 UTC" firstStartedPulling="2026-02-15 20:51:58.854805342 +0000 UTC m=+2126.720820965" lastFinishedPulling="2026-02-15 20:51:59.332748058 +0000 UTC m=+2127.198763681" observedRunningTime="2026-02-15 20:51:59.831303903 +0000 UTC m=+2127.697319526" watchObservedRunningTime="2026-02-15 20:51:59.834371435 +0000 UTC m=+2127.700387058" Feb 15 20:52:07 crc kubenswrapper[4735]: I0215 20:52:07.935679 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nw9vj"] Feb 15 20:52:07 crc kubenswrapper[4735]: I0215 20:52:07.938284 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:07 crc kubenswrapper[4735]: I0215 20:52:07.953419 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nw9vj"] Feb 15 20:52:07 crc kubenswrapper[4735]: I0215 20:52:07.990488 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skfk6\" (UniqueName: \"kubernetes.io/projected/d9509738-8192-470e-8761-03ec5e9589fd-kube-api-access-skfk6\") pod \"certified-operators-nw9vj\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:07 crc kubenswrapper[4735]: I0215 20:52:07.990773 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-catalog-content\") pod \"certified-operators-nw9vj\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:07 crc kubenswrapper[4735]: I0215 20:52:07.990801 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-utilities\") pod \"certified-operators-nw9vj\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:08 crc kubenswrapper[4735]: I0215 20:52:08.092271 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-utilities\") pod \"certified-operators-nw9vj\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:08 crc kubenswrapper[4735]: I0215 20:52:08.092380 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skfk6\" (UniqueName: \"kubernetes.io/projected/d9509738-8192-470e-8761-03ec5e9589fd-kube-api-access-skfk6\") pod \"certified-operators-nw9vj\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:08 crc kubenswrapper[4735]: I0215 20:52:08.092461 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-catalog-content\") pod \"certified-operators-nw9vj\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:08 crc kubenswrapper[4735]: I0215 20:52:08.092750 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-utilities\") pod \"certified-operators-nw9vj\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:08 crc kubenswrapper[4735]: I0215 20:52:08.092832 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-catalog-content\") pod \"certified-operators-nw9vj\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:08 crc kubenswrapper[4735]: I0215 20:52:08.120214 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skfk6\" (UniqueName: \"kubernetes.io/projected/d9509738-8192-470e-8761-03ec5e9589fd-kube-api-access-skfk6\") pod \"certified-operators-nw9vj\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:08 crc kubenswrapper[4735]: I0215 20:52:08.257331 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:08 crc kubenswrapper[4735]: I0215 20:52:08.744018 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nw9vj"] Feb 15 20:52:08 crc kubenswrapper[4735]: I0215 20:52:08.881926 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nw9vj" event={"ID":"d9509738-8192-470e-8761-03ec5e9589fd","Type":"ContainerStarted","Data":"e186326df82bd0876a134ec9e4adf2f0dd6668ce274ff328eb71a0664dfe6045"} Feb 15 20:52:09 crc kubenswrapper[4735]: I0215 20:52:09.891391 4735 generic.go:334] "Generic (PLEG): container finished" podID="d9509738-8192-470e-8761-03ec5e9589fd" containerID="53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7" exitCode=0 Feb 15 20:52:09 crc kubenswrapper[4735]: I0215 20:52:09.892025 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nw9vj" event={"ID":"d9509738-8192-470e-8761-03ec5e9589fd","Type":"ContainerDied","Data":"53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7"} Feb 15 20:52:10 crc kubenswrapper[4735]: I0215 20:52:10.904596 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nw9vj" event={"ID":"d9509738-8192-470e-8761-03ec5e9589fd","Type":"ContainerStarted","Data":"630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2"} Feb 15 20:52:11 crc kubenswrapper[4735]: I0215 20:52:11.962211 4735 generic.go:334] "Generic (PLEG): container finished" podID="d9509738-8192-470e-8761-03ec5e9589fd" containerID="630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2" exitCode=0 Feb 15 20:52:11 crc kubenswrapper[4735]: I0215 20:52:11.962284 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nw9vj" event={"ID":"d9509738-8192-470e-8761-03ec5e9589fd","Type":"ContainerDied","Data":"630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2"} Feb 15 20:52:12 crc kubenswrapper[4735]: I0215 20:52:12.971216 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nw9vj" event={"ID":"d9509738-8192-470e-8761-03ec5e9589fd","Type":"ContainerStarted","Data":"65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba"} Feb 15 20:52:12 crc kubenswrapper[4735]: I0215 20:52:12.991509 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nw9vj" podStartSLOduration=3.5307129699999997 podStartE2EDuration="5.991490514s" podCreationTimestamp="2026-02-15 20:52:07 +0000 UTC" firstStartedPulling="2026-02-15 20:52:09.893156379 +0000 UTC m=+2137.759172022" lastFinishedPulling="2026-02-15 20:52:12.353933943 +0000 UTC m=+2140.219949566" observedRunningTime="2026-02-15 20:52:12.987757255 +0000 UTC m=+2140.853772878" watchObservedRunningTime="2026-02-15 20:52:12.991490514 +0000 UTC m=+2140.857506137" Feb 15 20:52:18 crc kubenswrapper[4735]: I0215 20:52:18.258826 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:18 crc kubenswrapper[4735]: I0215 20:52:18.259371 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:18 crc kubenswrapper[4735]: I0215 20:52:18.309884 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:19 crc kubenswrapper[4735]: I0215 20:52:19.102581 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:19 crc kubenswrapper[4735]: I0215 20:52:19.182926 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nw9vj"] Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.028503 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nw9vj" podUID="d9509738-8192-470e-8761-03ec5e9589fd" containerName="registry-server" containerID="cri-o://65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba" gracePeriod=2 Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.571701 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.747929 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skfk6\" (UniqueName: \"kubernetes.io/projected/d9509738-8192-470e-8761-03ec5e9589fd-kube-api-access-skfk6\") pod \"d9509738-8192-470e-8761-03ec5e9589fd\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.748054 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-utilities\") pod \"d9509738-8192-470e-8761-03ec5e9589fd\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.748197 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-catalog-content\") pod \"d9509738-8192-470e-8761-03ec5e9589fd\" (UID: \"d9509738-8192-470e-8761-03ec5e9589fd\") " Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.748895 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-utilities" (OuterVolumeSpecName: "utilities") pod "d9509738-8192-470e-8761-03ec5e9589fd" (UID: "d9509738-8192-470e-8761-03ec5e9589fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.755715 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9509738-8192-470e-8761-03ec5e9589fd-kube-api-access-skfk6" (OuterVolumeSpecName: "kube-api-access-skfk6") pod "d9509738-8192-470e-8761-03ec5e9589fd" (UID: "d9509738-8192-470e-8761-03ec5e9589fd"). InnerVolumeSpecName "kube-api-access-skfk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.796025 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9509738-8192-470e-8761-03ec5e9589fd" (UID: "d9509738-8192-470e-8761-03ec5e9589fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.850081 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.850141 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skfk6\" (UniqueName: \"kubernetes.io/projected/d9509738-8192-470e-8761-03ec5e9589fd-kube-api-access-skfk6\") on node \"crc\" DevicePath \"\"" Feb 15 20:52:21 crc kubenswrapper[4735]: I0215 20:52:21.850163 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9509738-8192-470e-8761-03ec5e9589fd-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.039501 4735 generic.go:334] "Generic (PLEG): container finished" podID="d9509738-8192-470e-8761-03ec5e9589fd" containerID="65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba" exitCode=0 Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.039552 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nw9vj" event={"ID":"d9509738-8192-470e-8761-03ec5e9589fd","Type":"ContainerDied","Data":"65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba"} Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.039586 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nw9vj" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.039606 4735 scope.go:117] "RemoveContainer" containerID="65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.039588 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nw9vj" event={"ID":"d9509738-8192-470e-8761-03ec5e9589fd","Type":"ContainerDied","Data":"e186326df82bd0876a134ec9e4adf2f0dd6668ce274ff328eb71a0664dfe6045"} Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.069021 4735 scope.go:117] "RemoveContainer" containerID="630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.087484 4735 scope.go:117] "RemoveContainer" containerID="53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.097278 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nw9vj"] Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.107541 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nw9vj"] Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.155262 4735 scope.go:117] "RemoveContainer" containerID="65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba" Feb 15 20:52:22 crc kubenswrapper[4735]: E0215 20:52:22.158378 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba\": container with ID starting with 65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba not found: ID does not exist" containerID="65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.158418 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba"} err="failed to get container status \"65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba\": rpc error: code = NotFound desc = could not find container \"65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba\": container with ID starting with 65baa89931bbc963772fa6452a197966a3a3aea8472975f0261699d696146dba not found: ID does not exist" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.158462 4735 scope.go:117] "RemoveContainer" containerID="630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2" Feb 15 20:52:22 crc kubenswrapper[4735]: E0215 20:52:22.159035 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2\": container with ID starting with 630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2 not found: ID does not exist" containerID="630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.159071 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2"} err="failed to get container status \"630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2\": rpc error: code = NotFound desc = could not find container \"630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2\": container with ID starting with 630809e34e018321671302a30d56dd8c3180a25c382dc7785dcc26099e803ac2 not found: ID does not exist" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.159098 4735 scope.go:117] "RemoveContainer" containerID="53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7" Feb 15 20:52:22 crc kubenswrapper[4735]: E0215 20:52:22.159506 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7\": container with ID starting with 53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7 not found: ID does not exist" containerID="53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.159533 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7"} err="failed to get container status \"53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7\": rpc error: code = NotFound desc = could not find container \"53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7\": container with ID starting with 53fd27522db51e7d97f35a68e15dfcf6c8f8a656fd653a7ad51f290c3790f8c7 not found: ID does not exist" Feb 15 20:52:22 crc kubenswrapper[4735]: I0215 20:52:22.897621 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9509738-8192-470e-8761-03ec5e9589fd" path="/var/lib/kubelet/pods/d9509738-8192-470e-8761-03ec5e9589fd/volumes" Feb 15 20:53:32 crc kubenswrapper[4735]: I0215 20:53:32.966217 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tb98q"] Feb 15 20:53:32 crc kubenswrapper[4735]: E0215 20:53:32.967218 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9509738-8192-470e-8761-03ec5e9589fd" containerName="registry-server" Feb 15 20:53:32 crc kubenswrapper[4735]: I0215 20:53:32.967234 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9509738-8192-470e-8761-03ec5e9589fd" containerName="registry-server" Feb 15 20:53:32 crc kubenswrapper[4735]: E0215 20:53:32.967278 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9509738-8192-470e-8761-03ec5e9589fd" containerName="extract-content" Feb 15 20:53:32 crc kubenswrapper[4735]: I0215 20:53:32.967288 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9509738-8192-470e-8761-03ec5e9589fd" containerName="extract-content" Feb 15 20:53:32 crc kubenswrapper[4735]: E0215 20:53:32.967307 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9509738-8192-470e-8761-03ec5e9589fd" containerName="extract-utilities" Feb 15 20:53:32 crc kubenswrapper[4735]: I0215 20:53:32.967315 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9509738-8192-470e-8761-03ec5e9589fd" containerName="extract-utilities" Feb 15 20:53:32 crc kubenswrapper[4735]: I0215 20:53:32.967552 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9509738-8192-470e-8761-03ec5e9589fd" containerName="registry-server" Feb 15 20:53:32 crc kubenswrapper[4735]: I0215 20:53:32.969240 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:32 crc kubenswrapper[4735]: I0215 20:53:32.976123 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tb98q"] Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.136572 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96mmb\" (UniqueName: \"kubernetes.io/projected/24d81a28-bd21-43a1-8b25-fbe13082dd0b-kube-api-access-96mmb\") pod \"community-operators-tb98q\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.136709 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-catalog-content\") pod \"community-operators-tb98q\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.136794 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-utilities\") pod \"community-operators-tb98q\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.238307 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-catalog-content\") pod \"community-operators-tb98q\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.238672 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-utilities\") pod \"community-operators-tb98q\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.238845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96mmb\" (UniqueName: \"kubernetes.io/projected/24d81a28-bd21-43a1-8b25-fbe13082dd0b-kube-api-access-96mmb\") pod \"community-operators-tb98q\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.238987 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-catalog-content\") pod \"community-operators-tb98q\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.239122 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-utilities\") pod \"community-operators-tb98q\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.269209 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96mmb\" (UniqueName: \"kubernetes.io/projected/24d81a28-bd21-43a1-8b25-fbe13082dd0b-kube-api-access-96mmb\") pod \"community-operators-tb98q\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.302434 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:33 crc kubenswrapper[4735]: I0215 20:53:33.823147 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tb98q"] Feb 15 20:53:34 crc kubenswrapper[4735]: I0215 20:53:34.677233 4735 generic.go:334] "Generic (PLEG): container finished" podID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerID="10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f" exitCode=0 Feb 15 20:53:34 crc kubenswrapper[4735]: I0215 20:53:34.677314 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tb98q" event={"ID":"24d81a28-bd21-43a1-8b25-fbe13082dd0b","Type":"ContainerDied","Data":"10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f"} Feb 15 20:53:34 crc kubenswrapper[4735]: I0215 20:53:34.677564 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tb98q" event={"ID":"24d81a28-bd21-43a1-8b25-fbe13082dd0b","Type":"ContainerStarted","Data":"3a82e21d49012331240e99a78e830d632189293b5ac6b011df04bdea0cfb8531"} Feb 15 20:53:34 crc kubenswrapper[4735]: I0215 20:53:34.679352 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.160637 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tmplc"] Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.163139 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.173375 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmplc"] Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.283358 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-utilities\") pod \"redhat-marketplace-tmplc\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.283535 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z72m\" (UniqueName: \"kubernetes.io/projected/2e457536-dc5c-403b-8af6-ad450c0561f9-kube-api-access-7z72m\") pod \"redhat-marketplace-tmplc\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.283725 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-catalog-content\") pod \"redhat-marketplace-tmplc\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.385433 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z72m\" (UniqueName: \"kubernetes.io/projected/2e457536-dc5c-403b-8af6-ad450c0561f9-kube-api-access-7z72m\") pod \"redhat-marketplace-tmplc\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.385558 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-catalog-content\") pod \"redhat-marketplace-tmplc\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.385678 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-utilities\") pod \"redhat-marketplace-tmplc\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.386355 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-utilities\") pod \"redhat-marketplace-tmplc\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.386354 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-catalog-content\") pod \"redhat-marketplace-tmplc\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.405741 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z72m\" (UniqueName: \"kubernetes.io/projected/2e457536-dc5c-403b-8af6-ad450c0561f9-kube-api-access-7z72m\") pod \"redhat-marketplace-tmplc\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.481983 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.695185 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tb98q" event={"ID":"24d81a28-bd21-43a1-8b25-fbe13082dd0b","Type":"ContainerStarted","Data":"301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23"} Feb 15 20:53:35 crc kubenswrapper[4735]: I0215 20:53:35.988824 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmplc"] Feb 15 20:53:36 crc kubenswrapper[4735]: I0215 20:53:36.703852 4735 generic.go:334] "Generic (PLEG): container finished" podID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerID="21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5" exitCode=0 Feb 15 20:53:36 crc kubenswrapper[4735]: I0215 20:53:36.703890 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmplc" event={"ID":"2e457536-dc5c-403b-8af6-ad450c0561f9","Type":"ContainerDied","Data":"21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5"} Feb 15 20:53:36 crc kubenswrapper[4735]: I0215 20:53:36.704144 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmplc" event={"ID":"2e457536-dc5c-403b-8af6-ad450c0561f9","Type":"ContainerStarted","Data":"5f75fd5bc62dc32af4bae37f23a0e6ca47d2720020f0b908e65eefdd6fc57a04"} Feb 15 20:53:37 crc kubenswrapper[4735]: I0215 20:53:37.745303 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmplc" event={"ID":"2e457536-dc5c-403b-8af6-ad450c0561f9","Type":"ContainerStarted","Data":"4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32"} Feb 15 20:53:37 crc kubenswrapper[4735]: I0215 20:53:37.749757 4735 generic.go:334] "Generic (PLEG): container finished" podID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerID="301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23" exitCode=0 Feb 15 20:53:37 crc kubenswrapper[4735]: I0215 20:53:37.749824 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tb98q" event={"ID":"24d81a28-bd21-43a1-8b25-fbe13082dd0b","Type":"ContainerDied","Data":"301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23"} Feb 15 20:53:38 crc kubenswrapper[4735]: I0215 20:53:38.759598 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tb98q" event={"ID":"24d81a28-bd21-43a1-8b25-fbe13082dd0b","Type":"ContainerStarted","Data":"269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632"} Feb 15 20:53:38 crc kubenswrapper[4735]: I0215 20:53:38.762303 4735 generic.go:334] "Generic (PLEG): container finished" podID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerID="4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32" exitCode=0 Feb 15 20:53:38 crc kubenswrapper[4735]: I0215 20:53:38.762334 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmplc" event={"ID":"2e457536-dc5c-403b-8af6-ad450c0561f9","Type":"ContainerDied","Data":"4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32"} Feb 15 20:53:38 crc kubenswrapper[4735]: I0215 20:53:38.786767 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tb98q" podStartSLOduration=3.247046836 podStartE2EDuration="6.786743428s" podCreationTimestamp="2026-02-15 20:53:32 +0000 UTC" firstStartedPulling="2026-02-15 20:53:34.679129666 +0000 UTC m=+2222.545145289" lastFinishedPulling="2026-02-15 20:53:38.218826258 +0000 UTC m=+2226.084841881" observedRunningTime="2026-02-15 20:53:38.77861272 +0000 UTC m=+2226.644628363" watchObservedRunningTime="2026-02-15 20:53:38.786743428 +0000 UTC m=+2226.652759051" Feb 15 20:53:39 crc kubenswrapper[4735]: I0215 20:53:39.772428 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmplc" event={"ID":"2e457536-dc5c-403b-8af6-ad450c0561f9","Type":"ContainerStarted","Data":"41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509"} Feb 15 20:53:39 crc kubenswrapper[4735]: I0215 20:53:39.795308 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tmplc" podStartSLOduration=2.298137469 podStartE2EDuration="4.795290935s" podCreationTimestamp="2026-02-15 20:53:35 +0000 UTC" firstStartedPulling="2026-02-15 20:53:36.705659123 +0000 UTC m=+2224.571674746" lastFinishedPulling="2026-02-15 20:53:39.202812569 +0000 UTC m=+2227.068828212" observedRunningTime="2026-02-15 20:53:39.78724856 +0000 UTC m=+2227.653264203" watchObservedRunningTime="2026-02-15 20:53:39.795290935 +0000 UTC m=+2227.661306558" Feb 15 20:53:43 crc kubenswrapper[4735]: I0215 20:53:43.303546 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:43 crc kubenswrapper[4735]: I0215 20:53:43.303810 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:44 crc kubenswrapper[4735]: I0215 20:53:44.362280 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tb98q" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerName="registry-server" probeResult="failure" output=< Feb 15 20:53:44 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 20:53:44 crc kubenswrapper[4735]: > Feb 15 20:53:45 crc kubenswrapper[4735]: I0215 20:53:45.482484 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:45 crc kubenswrapper[4735]: I0215 20:53:45.482537 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:45 crc kubenswrapper[4735]: I0215 20:53:45.545097 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:45 crc kubenswrapper[4735]: I0215 20:53:45.880477 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:45 crc kubenswrapper[4735]: I0215 20:53:45.959858 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmplc"] Feb 15 20:53:47 crc kubenswrapper[4735]: I0215 20:53:47.848082 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tmplc" podUID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerName="registry-server" containerID="cri-o://41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509" gracePeriod=2 Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.316667 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.458714 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z72m\" (UniqueName: \"kubernetes.io/projected/2e457536-dc5c-403b-8af6-ad450c0561f9-kube-api-access-7z72m\") pod \"2e457536-dc5c-403b-8af6-ad450c0561f9\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.459216 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-catalog-content\") pod \"2e457536-dc5c-403b-8af6-ad450c0561f9\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.459382 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-utilities\") pod \"2e457536-dc5c-403b-8af6-ad450c0561f9\" (UID: \"2e457536-dc5c-403b-8af6-ad450c0561f9\") " Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.460089 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-utilities" (OuterVolumeSpecName: "utilities") pod "2e457536-dc5c-403b-8af6-ad450c0561f9" (UID: "2e457536-dc5c-403b-8af6-ad450c0561f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.473200 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e457536-dc5c-403b-8af6-ad450c0561f9-kube-api-access-7z72m" (OuterVolumeSpecName: "kube-api-access-7z72m") pod "2e457536-dc5c-403b-8af6-ad450c0561f9" (UID: "2e457536-dc5c-403b-8af6-ad450c0561f9"). InnerVolumeSpecName "kube-api-access-7z72m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.480644 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e457536-dc5c-403b-8af6-ad450c0561f9" (UID: "2e457536-dc5c-403b-8af6-ad450c0561f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.563299 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.563344 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z72m\" (UniqueName: \"kubernetes.io/projected/2e457536-dc5c-403b-8af6-ad450c0561f9-kube-api-access-7z72m\") on node \"crc\" DevicePath \"\"" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.563362 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e457536-dc5c-403b-8af6-ad450c0561f9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.855354 4735 generic.go:334] "Generic (PLEG): container finished" podID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerID="41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509" exitCode=0 Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.855411 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmplc" event={"ID":"2e457536-dc5c-403b-8af6-ad450c0561f9","Type":"ContainerDied","Data":"41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509"} Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.855469 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmplc" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.855498 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmplc" event={"ID":"2e457536-dc5c-403b-8af6-ad450c0561f9","Type":"ContainerDied","Data":"5f75fd5bc62dc32af4bae37f23a0e6ca47d2720020f0b908e65eefdd6fc57a04"} Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.855526 4735 scope.go:117] "RemoveContainer" containerID="41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.895806 4735 scope.go:117] "RemoveContainer" containerID="4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.899729 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmplc"] Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.914416 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmplc"] Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.920131 4735 scope.go:117] "RemoveContainer" containerID="21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.956005 4735 scope.go:117] "RemoveContainer" containerID="41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509" Feb 15 20:53:48 crc kubenswrapper[4735]: E0215 20:53:48.956354 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509\": container with ID starting with 41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509 not found: ID does not exist" containerID="41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.956377 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509"} err="failed to get container status \"41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509\": rpc error: code = NotFound desc = could not find container \"41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509\": container with ID starting with 41ad553248a47507e1fbf44d6facb84a3ffcd7e728652bcd770db81a1b20b509 not found: ID does not exist" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.956399 4735 scope.go:117] "RemoveContainer" containerID="4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32" Feb 15 20:53:48 crc kubenswrapper[4735]: E0215 20:53:48.956777 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32\": container with ID starting with 4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32 not found: ID does not exist" containerID="4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.956800 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32"} err="failed to get container status \"4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32\": rpc error: code = NotFound desc = could not find container \"4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32\": container with ID starting with 4f1fc6cbc77dab269356d735b356d74b03f015f509afda66eb8d3409be8b4d32 not found: ID does not exist" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.956813 4735 scope.go:117] "RemoveContainer" containerID="21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5" Feb 15 20:53:48 crc kubenswrapper[4735]: E0215 20:53:48.957175 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5\": container with ID starting with 21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5 not found: ID does not exist" containerID="21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5" Feb 15 20:53:48 crc kubenswrapper[4735]: I0215 20:53:48.957193 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5"} err="failed to get container status \"21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5\": rpc error: code = NotFound desc = could not find container \"21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5\": container with ID starting with 21c3a2b832f4aa974020407f8b95bee5f01f816b38af1d7c25e96b29a6a0d7d5 not found: ID does not exist" Feb 15 20:53:50 crc kubenswrapper[4735]: I0215 20:53:50.903261 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e457536-dc5c-403b-8af6-ad450c0561f9" path="/var/lib/kubelet/pods/2e457536-dc5c-403b-8af6-ad450c0561f9/volumes" Feb 15 20:53:53 crc kubenswrapper[4735]: I0215 20:53:53.349646 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:53 crc kubenswrapper[4735]: I0215 20:53:53.400193 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:53 crc kubenswrapper[4735]: I0215 20:53:53.591040 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tb98q"] Feb 15 20:53:54 crc kubenswrapper[4735]: I0215 20:53:54.924842 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tb98q" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerName="registry-server" containerID="cri-o://269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632" gracePeriod=2 Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.386824 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.509218 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-utilities\") pod \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.509471 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-catalog-content\") pod \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.509529 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96mmb\" (UniqueName: \"kubernetes.io/projected/24d81a28-bd21-43a1-8b25-fbe13082dd0b-kube-api-access-96mmb\") pod \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\" (UID: \"24d81a28-bd21-43a1-8b25-fbe13082dd0b\") " Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.510186 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-utilities" (OuterVolumeSpecName: "utilities") pod "24d81a28-bd21-43a1-8b25-fbe13082dd0b" (UID: "24d81a28-bd21-43a1-8b25-fbe13082dd0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.518338 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d81a28-bd21-43a1-8b25-fbe13082dd0b-kube-api-access-96mmb" (OuterVolumeSpecName: "kube-api-access-96mmb") pod "24d81a28-bd21-43a1-8b25-fbe13082dd0b" (UID: "24d81a28-bd21-43a1-8b25-fbe13082dd0b"). InnerVolumeSpecName "kube-api-access-96mmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.564429 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24d81a28-bd21-43a1-8b25-fbe13082dd0b" (UID: "24d81a28-bd21-43a1-8b25-fbe13082dd0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.611915 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.611981 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96mmb\" (UniqueName: \"kubernetes.io/projected/24d81a28-bd21-43a1-8b25-fbe13082dd0b-kube-api-access-96mmb\") on node \"crc\" DevicePath \"\"" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.611992 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d81a28-bd21-43a1-8b25-fbe13082dd0b-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.938905 4735 generic.go:334] "Generic (PLEG): container finished" podID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerID="269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632" exitCode=0 Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.939019 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tb98q" event={"ID":"24d81a28-bd21-43a1-8b25-fbe13082dd0b","Type":"ContainerDied","Data":"269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632"} Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.939048 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tb98q" event={"ID":"24d81a28-bd21-43a1-8b25-fbe13082dd0b","Type":"ContainerDied","Data":"3a82e21d49012331240e99a78e830d632189293b5ac6b011df04bdea0cfb8531"} Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.939056 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tb98q" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.939069 4735 scope.go:117] "RemoveContainer" containerID="269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.968333 4735 scope.go:117] "RemoveContainer" containerID="301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23" Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.984484 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tb98q"] Feb 15 20:53:55 crc kubenswrapper[4735]: I0215 20:53:55.992772 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tb98q"] Feb 15 20:53:56 crc kubenswrapper[4735]: I0215 20:53:56.000867 4735 scope.go:117] "RemoveContainer" containerID="10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f" Feb 15 20:53:56 crc kubenswrapper[4735]: I0215 20:53:56.040131 4735 scope.go:117] "RemoveContainer" containerID="269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632" Feb 15 20:53:56 crc kubenswrapper[4735]: E0215 20:53:56.040616 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632\": container with ID starting with 269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632 not found: ID does not exist" containerID="269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632" Feb 15 20:53:56 crc kubenswrapper[4735]: I0215 20:53:56.040722 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632"} err="failed to get container status \"269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632\": rpc error: code = NotFound desc = could not find container \"269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632\": container with ID starting with 269c6ccdf15bbe4baff186a1d5171c6109c181795445bf4983a9a2ca5be4e632 not found: ID does not exist" Feb 15 20:53:56 crc kubenswrapper[4735]: I0215 20:53:56.040835 4735 scope.go:117] "RemoveContainer" containerID="301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23" Feb 15 20:53:56 crc kubenswrapper[4735]: E0215 20:53:56.041297 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23\": container with ID starting with 301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23 not found: ID does not exist" containerID="301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23" Feb 15 20:53:56 crc kubenswrapper[4735]: I0215 20:53:56.041327 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23"} err="failed to get container status \"301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23\": rpc error: code = NotFound desc = could not find container \"301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23\": container with ID starting with 301858703a970ed72925f56f05029b22c519617900b81cc4dd3b611a7455bb23 not found: ID does not exist" Feb 15 20:53:56 crc kubenswrapper[4735]: I0215 20:53:56.041349 4735 scope.go:117] "RemoveContainer" containerID="10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f" Feb 15 20:53:56 crc kubenswrapper[4735]: E0215 20:53:56.041683 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f\": container with ID starting with 10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f not found: ID does not exist" containerID="10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f" Feb 15 20:53:56 crc kubenswrapper[4735]: I0215 20:53:56.041790 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f"} err="failed to get container status \"10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f\": rpc error: code = NotFound desc = could not find container \"10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f\": container with ID starting with 10a1e81aef39f0372448383a98802c280ea12e138acde8728de02a594beb1b1f not found: ID does not exist" Feb 15 20:53:56 crc kubenswrapper[4735]: I0215 20:53:56.944126 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" path="/var/lib/kubelet/pods/24d81a28-bd21-43a1-8b25-fbe13082dd0b/volumes" Feb 15 20:54:19 crc kubenswrapper[4735]: I0215 20:54:19.680312 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:54:19 crc kubenswrapper[4735]: I0215 20:54:19.682201 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:54:49 crc kubenswrapper[4735]: I0215 20:54:49.679438 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:54:49 crc kubenswrapper[4735]: I0215 20:54:49.679899 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:55:19 crc kubenswrapper[4735]: I0215 20:55:19.680397 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 20:55:19 crc kubenswrapper[4735]: I0215 20:55:19.682311 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 20:55:19 crc kubenswrapper[4735]: I0215 20:55:19.682441 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 20:55:19 crc kubenswrapper[4735]: I0215 20:55:19.683179 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 20:55:19 crc kubenswrapper[4735]: I0215 20:55:19.683231 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" gracePeriod=600 Feb 15 20:55:19 crc kubenswrapper[4735]: E0215 20:55:19.810738 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:55:20 crc kubenswrapper[4735]: I0215 20:55:20.746693 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" exitCode=0 Feb 15 20:55:20 crc kubenswrapper[4735]: I0215 20:55:20.746782 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639"} Feb 15 20:55:20 crc kubenswrapper[4735]: I0215 20:55:20.747062 4735 scope.go:117] "RemoveContainer" containerID="728ffce9a88b0b401f4b8e2c49fe36be9cdd7e5f2f6a5c64dec18e527e842d18" Feb 15 20:55:20 crc kubenswrapper[4735]: I0215 20:55:20.747882 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:55:20 crc kubenswrapper[4735]: E0215 20:55:20.748757 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:55:34 crc kubenswrapper[4735]: I0215 20:55:34.887826 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:55:34 crc kubenswrapper[4735]: E0215 20:55:34.888871 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:55:48 crc kubenswrapper[4735]: I0215 20:55:48.886484 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:55:48 crc kubenswrapper[4735]: E0215 20:55:48.887234 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:56:00 crc kubenswrapper[4735]: I0215 20:56:00.886992 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:56:00 crc kubenswrapper[4735]: E0215 20:56:00.887794 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:56:11 crc kubenswrapper[4735]: I0215 20:56:11.886744 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:56:11 crc kubenswrapper[4735]: E0215 20:56:11.887392 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:56:14 crc kubenswrapper[4735]: I0215 20:56:14.238897 4735 generic.go:334] "Generic (PLEG): container finished" podID="3e63486b-442e-477f-8692-cb3da5e648cd" containerID="ce0ecdb4914f3d103027f9eaacc71fe6b9173e1d899efb53ff0134ac79a6e70d" exitCode=0 Feb 15 20:56:14 crc kubenswrapper[4735]: I0215 20:56:14.239208 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" event={"ID":"3e63486b-442e-477f-8692-cb3da5e648cd","Type":"ContainerDied","Data":"ce0ecdb4914f3d103027f9eaacc71fe6b9173e1d899efb53ff0134ac79a6e70d"} Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.692072 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.893138 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-ssh-key-openstack-edpm-ipam\") pod \"3e63486b-442e-477f-8692-cb3da5e648cd\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.893498 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-secret-0\") pod \"3e63486b-442e-477f-8692-cb3da5e648cd\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.893740 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-combined-ca-bundle\") pod \"3e63486b-442e-477f-8692-cb3da5e648cd\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.893787 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-inventory\") pod \"3e63486b-442e-477f-8692-cb3da5e648cd\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.893880 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5wjl\" (UniqueName: \"kubernetes.io/projected/3e63486b-442e-477f-8692-cb3da5e648cd-kube-api-access-d5wjl\") pod \"3e63486b-442e-477f-8692-cb3da5e648cd\" (UID: \"3e63486b-442e-477f-8692-cb3da5e648cd\") " Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.920994 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e63486b-442e-477f-8692-cb3da5e648cd-kube-api-access-d5wjl" (OuterVolumeSpecName: "kube-api-access-d5wjl") pod "3e63486b-442e-477f-8692-cb3da5e648cd" (UID: "3e63486b-442e-477f-8692-cb3da5e648cd"). InnerVolumeSpecName "kube-api-access-d5wjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.923876 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3e63486b-442e-477f-8692-cb3da5e648cd" (UID: "3e63486b-442e-477f-8692-cb3da5e648cd"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.953278 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3e63486b-442e-477f-8692-cb3da5e648cd" (UID: "3e63486b-442e-477f-8692-cb3da5e648cd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.955689 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-inventory" (OuterVolumeSpecName: "inventory") pod "3e63486b-442e-477f-8692-cb3da5e648cd" (UID: "3e63486b-442e-477f-8692-cb3da5e648cd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.956914 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "3e63486b-442e-477f-8692-cb3da5e648cd" (UID: "3e63486b-442e-477f-8692-cb3da5e648cd"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.996450 4735 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.996818 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.996848 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5wjl\" (UniqueName: \"kubernetes.io/projected/3e63486b-442e-477f-8692-cb3da5e648cd-kube-api-access-d5wjl\") on node \"crc\" DevicePath \"\"" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.996862 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:56:15 crc kubenswrapper[4735]: I0215 20:56:15.996874 4735 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3e63486b-442e-477f-8692-cb3da5e648cd-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.259446 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" event={"ID":"3e63486b-442e-477f-8692-cb3da5e648cd","Type":"ContainerDied","Data":"323209fbcba95da9c5a2645693181656b5fd2011a6db912fd0290164a7e5b312"} Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.259484 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="323209fbcba95da9c5a2645693181656b5fd2011a6db912fd0290164a7e5b312" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.259742 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.366976 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42"] Feb 15 20:56:16 crc kubenswrapper[4735]: E0215 20:56:16.367424 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerName="extract-utilities" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367446 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerName="extract-utilities" Feb 15 20:56:16 crc kubenswrapper[4735]: E0215 20:56:16.367468 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerName="extract-utilities" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367477 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerName="extract-utilities" Feb 15 20:56:16 crc kubenswrapper[4735]: E0215 20:56:16.367496 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerName="extract-content" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367505 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerName="extract-content" Feb 15 20:56:16 crc kubenswrapper[4735]: E0215 20:56:16.367524 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerName="extract-content" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367532 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerName="extract-content" Feb 15 20:56:16 crc kubenswrapper[4735]: E0215 20:56:16.367545 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerName="registry-server" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367554 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerName="registry-server" Feb 15 20:56:16 crc kubenswrapper[4735]: E0215 20:56:16.367569 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerName="registry-server" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367578 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerName="registry-server" Feb 15 20:56:16 crc kubenswrapper[4735]: E0215 20:56:16.367589 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e63486b-442e-477f-8692-cb3da5e648cd" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367598 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e63486b-442e-477f-8692-cb3da5e648cd" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367808 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e457536-dc5c-403b-8af6-ad450c0561f9" containerName="registry-server" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367837 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d81a28-bd21-43a1-8b25-fbe13082dd0b" containerName="registry-server" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.367881 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e63486b-442e-477f-8692-cb3da5e648cd" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.368715 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.371243 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.371464 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.371672 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.372354 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.372445 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.372483 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.372593 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.394110 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42"] Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.512529 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.512592 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.512622 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfxxm\" (UniqueName: \"kubernetes.io/projected/01424332-5cb6-4959-90b3-eb98af1ee96e-kube-api-access-tfxxm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.512687 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.512714 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.512735 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.512753 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.512971 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.513042 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.513147 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.513210 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.614877 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.615022 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.615075 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfxxm\" (UniqueName: \"kubernetes.io/projected/01424332-5cb6-4959-90b3-eb98af1ee96e-kube-api-access-tfxxm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.615191 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.615240 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.615282 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.615315 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.615384 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.616812 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.616902 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.617001 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.618000 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.623782 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.623988 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.624329 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.625016 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.625839 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.626355 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.626590 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.631131 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.640789 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.645782 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfxxm\" (UniqueName: \"kubernetes.io/projected/01424332-5cb6-4959-90b3-eb98af1ee96e-kube-api-access-tfxxm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t2s42\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:16 crc kubenswrapper[4735]: I0215 20:56:16.691842 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:56:17 crc kubenswrapper[4735]: I0215 20:56:17.255694 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42"] Feb 15 20:56:17 crc kubenswrapper[4735]: I0215 20:56:17.269748 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" event={"ID":"01424332-5cb6-4959-90b3-eb98af1ee96e","Type":"ContainerStarted","Data":"d19eea7cd259d2826099c3f4da2faaeefbd3cacac1e9dc4438c43bbf3088816b"} Feb 15 20:56:18 crc kubenswrapper[4735]: I0215 20:56:18.278784 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" event={"ID":"01424332-5cb6-4959-90b3-eb98af1ee96e","Type":"ContainerStarted","Data":"c488bb29892e84c47b3063a815f2d64e6262dd5232c98fa2104ca7b2cf7bae97"} Feb 15 20:56:18 crc kubenswrapper[4735]: I0215 20:56:18.312031 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" podStartSLOduration=1.847153184 podStartE2EDuration="2.312015609s" podCreationTimestamp="2026-02-15 20:56:16 +0000 UTC" firstStartedPulling="2026-02-15 20:56:17.252302394 +0000 UTC m=+2385.118318017" lastFinishedPulling="2026-02-15 20:56:17.717164819 +0000 UTC m=+2385.583180442" observedRunningTime="2026-02-15 20:56:18.302880675 +0000 UTC m=+2386.168896298" watchObservedRunningTime="2026-02-15 20:56:18.312015609 +0000 UTC m=+2386.178031222" Feb 15 20:56:26 crc kubenswrapper[4735]: I0215 20:56:26.887042 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:56:26 crc kubenswrapper[4735]: E0215 20:56:26.887963 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:56:38 crc kubenswrapper[4735]: I0215 20:56:38.887768 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:56:38 crc kubenswrapper[4735]: E0215 20:56:38.888610 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:56:52 crc kubenswrapper[4735]: I0215 20:56:52.897127 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:56:52 crc kubenswrapper[4735]: E0215 20:56:52.898151 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:57:06 crc kubenswrapper[4735]: I0215 20:57:06.886379 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:57:06 crc kubenswrapper[4735]: E0215 20:57:06.888067 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:57:17 crc kubenswrapper[4735]: I0215 20:57:17.886331 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:57:17 crc kubenswrapper[4735]: E0215 20:57:17.887140 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:57:32 crc kubenswrapper[4735]: I0215 20:57:32.900862 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:57:32 crc kubenswrapper[4735]: E0215 20:57:32.901726 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:57:44 crc kubenswrapper[4735]: I0215 20:57:44.886695 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:57:44 crc kubenswrapper[4735]: E0215 20:57:44.887455 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:57:58 crc kubenswrapper[4735]: I0215 20:57:58.890239 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:57:58 crc kubenswrapper[4735]: E0215 20:57:58.891477 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:58:11 crc kubenswrapper[4735]: I0215 20:58:11.887327 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:58:11 crc kubenswrapper[4735]: E0215 20:58:11.889161 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:58:26 crc kubenswrapper[4735]: I0215 20:58:26.887692 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:58:26 crc kubenswrapper[4735]: E0215 20:58:26.888840 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:58:38 crc kubenswrapper[4735]: I0215 20:58:38.886967 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:58:38 crc kubenswrapper[4735]: E0215 20:58:38.890215 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:58:51 crc kubenswrapper[4735]: I0215 20:58:51.886745 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:58:51 crc kubenswrapper[4735]: E0215 20:58:51.887594 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:59:02 crc kubenswrapper[4735]: I0215 20:59:02.893034 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:59:02 crc kubenswrapper[4735]: E0215 20:59:02.893918 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:59:04 crc kubenswrapper[4735]: I0215 20:59:04.829982 4735 generic.go:334] "Generic (PLEG): container finished" podID="01424332-5cb6-4959-90b3-eb98af1ee96e" containerID="c488bb29892e84c47b3063a815f2d64e6262dd5232c98fa2104ca7b2cf7bae97" exitCode=0 Feb 15 20:59:04 crc kubenswrapper[4735]: I0215 20:59:04.830174 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" event={"ID":"01424332-5cb6-4959-90b3-eb98af1ee96e","Type":"ContainerDied","Data":"c488bb29892e84c47b3063a815f2d64e6262dd5232c98fa2104ca7b2cf7bae97"} Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.231293 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.392290 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-1\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.392749 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-inventory\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.392821 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfxxm\" (UniqueName: \"kubernetes.io/projected/01424332-5cb6-4959-90b3-eb98af1ee96e-kube-api-access-tfxxm\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.392870 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-extra-config-0\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.392980 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-0\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.393082 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-1\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.393129 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-combined-ca-bundle\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.393226 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-0\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.393274 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-ssh-key-openstack-edpm-ipam\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.393295 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-2\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.393315 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-3\") pod \"01424332-5cb6-4959-90b3-eb98af1ee96e\" (UID: \"01424332-5cb6-4959-90b3-eb98af1ee96e\") " Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.429873 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01424332-5cb6-4959-90b3-eb98af1ee96e-kube-api-access-tfxxm" (OuterVolumeSpecName: "kube-api-access-tfxxm") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "kube-api-access-tfxxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.431094 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.437156 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.440431 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.454342 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.456309 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.458031 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.468514 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-inventory" (OuterVolumeSpecName: "inventory") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.469894 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.471655 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.473520 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "01424332-5cb6-4959-90b3-eb98af1ee96e" (UID: "01424332-5cb6-4959-90b3-eb98af1ee96e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.497968 4735 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498207 4735 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498285 4735 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498442 4735 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498514 4735 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498570 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498632 4735 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498690 4735 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498752 4735 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498808 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01424332-5cb6-4959-90b3-eb98af1ee96e-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.498865 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfxxm\" (UniqueName: \"kubernetes.io/projected/01424332-5cb6-4959-90b3-eb98af1ee96e-kube-api-access-tfxxm\") on node \"crc\" DevicePath \"\"" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.854423 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" event={"ID":"01424332-5cb6-4959-90b3-eb98af1ee96e","Type":"ContainerDied","Data":"d19eea7cd259d2826099c3f4da2faaeefbd3cacac1e9dc4438c43bbf3088816b"} Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.854469 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d19eea7cd259d2826099c3f4da2faaeefbd3cacac1e9dc4438c43bbf3088816b" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.854472 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t2s42" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.946144 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf"] Feb 15 20:59:06 crc kubenswrapper[4735]: E0215 20:59:06.946775 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01424332-5cb6-4959-90b3-eb98af1ee96e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.946846 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="01424332-5cb6-4959-90b3-eb98af1ee96e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.947096 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="01424332-5cb6-4959-90b3-eb98af1ee96e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.947736 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.950437 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jhtck" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.960731 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf"] Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.961438 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.962433 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.962623 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 15 20:59:06 crc kubenswrapper[4735]: I0215 20:59:06.962817 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.109810 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.109898 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.109925 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.109973 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.110000 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbq5x\" (UniqueName: \"kubernetes.io/projected/25a254b5-abc4-45fa-aba5-935d79ae8959-kube-api-access-gbq5x\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.110261 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.110312 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.211651 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.211695 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.211717 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.211797 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.211819 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.211845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.211865 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbq5x\" (UniqueName: \"kubernetes.io/projected/25a254b5-abc4-45fa-aba5-935d79ae8959-kube-api-access-gbq5x\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.215887 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.216313 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.216656 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.219310 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.219389 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.225424 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.237310 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbq5x\" (UniqueName: \"kubernetes.io/projected/25a254b5-abc4-45fa-aba5-935d79ae8959-kube-api-access-gbq5x\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.274865 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.919109 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf"] Feb 15 20:59:07 crc kubenswrapper[4735]: I0215 20:59:07.923124 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 20:59:08 crc kubenswrapper[4735]: I0215 20:59:08.877841 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" event={"ID":"25a254b5-abc4-45fa-aba5-935d79ae8959","Type":"ContainerStarted","Data":"1c4da665bc4d39996eff53f8e95fc779c1dce1aba8c1062e0a464a10b8486c32"} Feb 15 20:59:08 crc kubenswrapper[4735]: I0215 20:59:08.878958 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" event={"ID":"25a254b5-abc4-45fa-aba5-935d79ae8959","Type":"ContainerStarted","Data":"0fbe16e878659bc737992b815754348b204c7f56d0c44aa60333a1e34f367ce2"} Feb 15 20:59:08 crc kubenswrapper[4735]: I0215 20:59:08.905188 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" podStartSLOduration=2.339193432 podStartE2EDuration="2.905168811s" podCreationTimestamp="2026-02-15 20:59:06 +0000 UTC" firstStartedPulling="2026-02-15 20:59:07.922799943 +0000 UTC m=+2555.788815556" lastFinishedPulling="2026-02-15 20:59:08.488775302 +0000 UTC m=+2556.354790935" observedRunningTime="2026-02-15 20:59:08.900058694 +0000 UTC m=+2556.766074307" watchObservedRunningTime="2026-02-15 20:59:08.905168811 +0000 UTC m=+2556.771184434" Feb 15 20:59:16 crc kubenswrapper[4735]: I0215 20:59:16.061117 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:59:16 crc kubenswrapper[4735]: E0215 20:59:16.063677 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:59:28 crc kubenswrapper[4735]: I0215 20:59:28.887709 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:59:28 crc kubenswrapper[4735]: E0215 20:59:28.888692 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:59:41 crc kubenswrapper[4735]: I0215 20:59:41.886889 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:59:41 crc kubenswrapper[4735]: E0215 20:59:41.887644 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 20:59:52 crc kubenswrapper[4735]: I0215 20:59:52.900838 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 20:59:52 crc kubenswrapper[4735]: E0215 20:59:52.901891 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.147436 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx"] Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.149093 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.151798 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.152058 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.209249 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx"] Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.234435 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sthlw\" (UniqueName: \"kubernetes.io/projected/16dfc773-d909-49eb-a20d-73ecf0a90271-kube-api-access-sthlw\") pod \"collect-profiles-29519820-kfshx\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.234505 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16dfc773-d909-49eb-a20d-73ecf0a90271-secret-volume\") pod \"collect-profiles-29519820-kfshx\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.234529 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16dfc773-d909-49eb-a20d-73ecf0a90271-config-volume\") pod \"collect-profiles-29519820-kfshx\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.336546 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sthlw\" (UniqueName: \"kubernetes.io/projected/16dfc773-d909-49eb-a20d-73ecf0a90271-kube-api-access-sthlw\") pod \"collect-profiles-29519820-kfshx\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.336629 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16dfc773-d909-49eb-a20d-73ecf0a90271-secret-volume\") pod \"collect-profiles-29519820-kfshx\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.336660 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16dfc773-d909-49eb-a20d-73ecf0a90271-config-volume\") pod \"collect-profiles-29519820-kfshx\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.337729 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16dfc773-d909-49eb-a20d-73ecf0a90271-config-volume\") pod \"collect-profiles-29519820-kfshx\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.343856 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16dfc773-d909-49eb-a20d-73ecf0a90271-secret-volume\") pod \"collect-profiles-29519820-kfshx\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.356756 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sthlw\" (UniqueName: \"kubernetes.io/projected/16dfc773-d909-49eb-a20d-73ecf0a90271-kube-api-access-sthlw\") pod \"collect-profiles-29519820-kfshx\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:00 crc kubenswrapper[4735]: I0215 21:00:00.522382 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:01 crc kubenswrapper[4735]: I0215 21:00:01.001901 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx"] Feb 15 21:00:01 crc kubenswrapper[4735]: I0215 21:00:01.607355 4735 generic.go:334] "Generic (PLEG): container finished" podID="16dfc773-d909-49eb-a20d-73ecf0a90271" containerID="461dc2b8e559b02c9d6e7ac3a52c974d8506c855c040c573c864b6a915de18cd" exitCode=0 Feb 15 21:00:01 crc kubenswrapper[4735]: I0215 21:00:01.607441 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" event={"ID":"16dfc773-d909-49eb-a20d-73ecf0a90271","Type":"ContainerDied","Data":"461dc2b8e559b02c9d6e7ac3a52c974d8506c855c040c573c864b6a915de18cd"} Feb 15 21:00:01 crc kubenswrapper[4735]: I0215 21:00:01.610522 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" event={"ID":"16dfc773-d909-49eb-a20d-73ecf0a90271","Type":"ContainerStarted","Data":"e3f78038465f233ade06bf9171888e0742f46313c74b88b9bc0cf679e43953d9"} Feb 15 21:00:02 crc kubenswrapper[4735]: I0215 21:00:02.999147 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.101441 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sthlw\" (UniqueName: \"kubernetes.io/projected/16dfc773-d909-49eb-a20d-73ecf0a90271-kube-api-access-sthlw\") pod \"16dfc773-d909-49eb-a20d-73ecf0a90271\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.101752 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16dfc773-d909-49eb-a20d-73ecf0a90271-secret-volume\") pod \"16dfc773-d909-49eb-a20d-73ecf0a90271\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.102424 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16dfc773-d909-49eb-a20d-73ecf0a90271-config-volume\") pod \"16dfc773-d909-49eb-a20d-73ecf0a90271\" (UID: \"16dfc773-d909-49eb-a20d-73ecf0a90271\") " Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.103309 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16dfc773-d909-49eb-a20d-73ecf0a90271-config-volume" (OuterVolumeSpecName: "config-volume") pod "16dfc773-d909-49eb-a20d-73ecf0a90271" (UID: "16dfc773-d909-49eb-a20d-73ecf0a90271"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.107420 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16dfc773-d909-49eb-a20d-73ecf0a90271-kube-api-access-sthlw" (OuterVolumeSpecName: "kube-api-access-sthlw") pod "16dfc773-d909-49eb-a20d-73ecf0a90271" (UID: "16dfc773-d909-49eb-a20d-73ecf0a90271"). InnerVolumeSpecName "kube-api-access-sthlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.111188 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16dfc773-d909-49eb-a20d-73ecf0a90271-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "16dfc773-d909-49eb-a20d-73ecf0a90271" (UID: "16dfc773-d909-49eb-a20d-73ecf0a90271"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.205431 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16dfc773-d909-49eb-a20d-73ecf0a90271-config-volume\") on node \"crc\" DevicePath \"\"" Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.205477 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sthlw\" (UniqueName: \"kubernetes.io/projected/16dfc773-d909-49eb-a20d-73ecf0a90271-kube-api-access-sthlw\") on node \"crc\" DevicePath \"\"" Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.205493 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16dfc773-d909-49eb-a20d-73ecf0a90271-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.628700 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" event={"ID":"16dfc773-d909-49eb-a20d-73ecf0a90271","Type":"ContainerDied","Data":"e3f78038465f233ade06bf9171888e0742f46313c74b88b9bc0cf679e43953d9"} Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.629094 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3f78038465f233ade06bf9171888e0742f46313c74b88b9bc0cf679e43953d9" Feb 15 21:00:03 crc kubenswrapper[4735]: I0215 21:00:03.628738 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx" Feb 15 21:00:04 crc kubenswrapper[4735]: I0215 21:00:04.082560 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h"] Feb 15 21:00:04 crc kubenswrapper[4735]: I0215 21:00:04.091815 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519775-4dm4h"] Feb 15 21:00:04 crc kubenswrapper[4735]: I0215 21:00:04.898250 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a930872e-2679-46c9-ac6f-22bffe1b8d66" path="/var/lib/kubelet/pods/a930872e-2679-46c9-ac6f-22bffe1b8d66/volumes" Feb 15 21:00:05 crc kubenswrapper[4735]: I0215 21:00:05.888160 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 21:00:05 crc kubenswrapper[4735]: E0215 21:00:05.888681 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:00:17 crc kubenswrapper[4735]: I0215 21:00:17.886868 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 21:00:17 crc kubenswrapper[4735]: E0215 21:00:17.887677 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:00:28 crc kubenswrapper[4735]: I0215 21:00:28.887155 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 21:00:29 crc kubenswrapper[4735]: I0215 21:00:29.871642 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"ad3865af6e8d595faa44438b03b1d22c4d0d2868c6797e3edfc5272ac148174c"} Feb 15 21:00:39 crc kubenswrapper[4735]: I0215 21:00:39.250581 4735 scope.go:117] "RemoveContainer" containerID="fd2feb5ae65c846e3bfc78616e103db5f08dd6d485bc8e98f0fea784efdca889" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.107280 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5gxdj"] Feb 15 21:00:49 crc kubenswrapper[4735]: E0215 21:00:49.108370 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16dfc773-d909-49eb-a20d-73ecf0a90271" containerName="collect-profiles" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.108391 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="16dfc773-d909-49eb-a20d-73ecf0a90271" containerName="collect-profiles" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.108640 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="16dfc773-d909-49eb-a20d-73ecf0a90271" containerName="collect-profiles" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.110422 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.168965 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5gxdj"] Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.209928 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwzzd\" (UniqueName: \"kubernetes.io/projected/94047cad-d62c-418d-aef3-a117a51a7ac9-kube-api-access-jwzzd\") pod \"redhat-operators-5gxdj\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.210054 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-catalog-content\") pod \"redhat-operators-5gxdj\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.210242 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-utilities\") pod \"redhat-operators-5gxdj\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.311607 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-catalog-content\") pod \"redhat-operators-5gxdj\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.311669 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-utilities\") pod \"redhat-operators-5gxdj\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.311751 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwzzd\" (UniqueName: \"kubernetes.io/projected/94047cad-d62c-418d-aef3-a117a51a7ac9-kube-api-access-jwzzd\") pod \"redhat-operators-5gxdj\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.312286 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-catalog-content\") pod \"redhat-operators-5gxdj\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.312587 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-utilities\") pod \"redhat-operators-5gxdj\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.332145 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwzzd\" (UniqueName: \"kubernetes.io/projected/94047cad-d62c-418d-aef3-a117a51a7ac9-kube-api-access-jwzzd\") pod \"redhat-operators-5gxdj\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.438703 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:49 crc kubenswrapper[4735]: I0215 21:00:49.941693 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5gxdj"] Feb 15 21:00:50 crc kubenswrapper[4735]: I0215 21:00:50.036789 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gxdj" event={"ID":"94047cad-d62c-418d-aef3-a117a51a7ac9","Type":"ContainerStarted","Data":"dcbdcf5171937303a8713654f3eb654ed8e462bbaecddb613c4bbe5012643cf4"} Feb 15 21:00:51 crc kubenswrapper[4735]: I0215 21:00:51.046187 4735 generic.go:334] "Generic (PLEG): container finished" podID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerID="59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9" exitCode=0 Feb 15 21:00:51 crc kubenswrapper[4735]: I0215 21:00:51.046277 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gxdj" event={"ID":"94047cad-d62c-418d-aef3-a117a51a7ac9","Type":"ContainerDied","Data":"59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9"} Feb 15 21:00:52 crc kubenswrapper[4735]: I0215 21:00:52.059208 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gxdj" event={"ID":"94047cad-d62c-418d-aef3-a117a51a7ac9","Type":"ContainerStarted","Data":"25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92"} Feb 15 21:00:57 crc kubenswrapper[4735]: I0215 21:00:57.104907 4735 generic.go:334] "Generic (PLEG): container finished" podID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerID="25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92" exitCode=0 Feb 15 21:00:57 crc kubenswrapper[4735]: I0215 21:00:57.105031 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gxdj" event={"ID":"94047cad-d62c-418d-aef3-a117a51a7ac9","Type":"ContainerDied","Data":"25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92"} Feb 15 21:00:58 crc kubenswrapper[4735]: I0215 21:00:58.114753 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gxdj" event={"ID":"94047cad-d62c-418d-aef3-a117a51a7ac9","Type":"ContainerStarted","Data":"022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81"} Feb 15 21:00:58 crc kubenswrapper[4735]: I0215 21:00:58.141360 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5gxdj" podStartSLOduration=2.6675794870000002 podStartE2EDuration="9.141336255s" podCreationTimestamp="2026-02-15 21:00:49 +0000 UTC" firstStartedPulling="2026-02-15 21:00:51.050047271 +0000 UTC m=+2658.916062894" lastFinishedPulling="2026-02-15 21:00:57.523803999 +0000 UTC m=+2665.389819662" observedRunningTime="2026-02-15 21:00:58.134123282 +0000 UTC m=+2666.000138905" watchObservedRunningTime="2026-02-15 21:00:58.141336255 +0000 UTC m=+2666.007351878" Feb 15 21:00:59 crc kubenswrapper[4735]: I0215 21:00:59.439579 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:00:59 crc kubenswrapper[4735]: I0215 21:00:59.439868 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.153859 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29519821-8vcgr"] Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.155074 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.179440 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29519821-8vcgr"] Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.308123 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv89r\" (UniqueName: \"kubernetes.io/projected/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-kube-api-access-gv89r\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.308687 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-fernet-keys\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.308761 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-config-data\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.309170 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-combined-ca-bundle\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.411119 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv89r\" (UniqueName: \"kubernetes.io/projected/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-kube-api-access-gv89r\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.411223 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-fernet-keys\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.411271 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-config-data\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.411375 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-combined-ca-bundle\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.417693 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-fernet-keys\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.418759 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-config-data\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.425857 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-combined-ca-bundle\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.439075 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv89r\" (UniqueName: \"kubernetes.io/projected/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-kube-api-access-gv89r\") pod \"keystone-cron-29519821-8vcgr\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.484247 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5gxdj" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="registry-server" probeResult="failure" output=< Feb 15 21:01:00 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:01:00 crc kubenswrapper[4735]: > Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.486305 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:00 crc kubenswrapper[4735]: I0215 21:01:00.994565 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29519821-8vcgr"] Feb 15 21:01:01 crc kubenswrapper[4735]: W0215 21:01:01.009152 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25286b3f_a5eb_49ce_a0dc_e2a8e99d9264.slice/crio-5f399ec52e45d02969d0d17bab7f85492a779bfffa96460d80fdda2f9ab76946 WatchSource:0}: Error finding container 5f399ec52e45d02969d0d17bab7f85492a779bfffa96460d80fdda2f9ab76946: Status 404 returned error can't find the container with id 5f399ec52e45d02969d0d17bab7f85492a779bfffa96460d80fdda2f9ab76946 Feb 15 21:01:01 crc kubenswrapper[4735]: I0215 21:01:01.140568 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29519821-8vcgr" event={"ID":"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264","Type":"ContainerStarted","Data":"5f399ec52e45d02969d0d17bab7f85492a779bfffa96460d80fdda2f9ab76946"} Feb 15 21:01:02 crc kubenswrapper[4735]: I0215 21:01:02.150665 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29519821-8vcgr" event={"ID":"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264","Type":"ContainerStarted","Data":"4635b35323ea05476d3e939b4e0d39e1bcef5d90cc2f18bba5e4f9d6211283a8"} Feb 15 21:01:02 crc kubenswrapper[4735]: I0215 21:01:02.178572 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29519821-8vcgr" podStartSLOduration=2.178552145 podStartE2EDuration="2.178552145s" podCreationTimestamp="2026-02-15 21:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 21:01:02.167250093 +0000 UTC m=+2670.033265716" watchObservedRunningTime="2026-02-15 21:01:02.178552145 +0000 UTC m=+2670.044567788" Feb 15 21:01:05 crc kubenswrapper[4735]: I0215 21:01:05.177841 4735 generic.go:334] "Generic (PLEG): container finished" podID="25286b3f-a5eb-49ce-a0dc-e2a8e99d9264" containerID="4635b35323ea05476d3e939b4e0d39e1bcef5d90cc2f18bba5e4f9d6211283a8" exitCode=0 Feb 15 21:01:05 crc kubenswrapper[4735]: I0215 21:01:05.177927 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29519821-8vcgr" event={"ID":"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264","Type":"ContainerDied","Data":"4635b35323ea05476d3e939b4e0d39e1bcef5d90cc2f18bba5e4f9d6211283a8"} Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.513668 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.584024 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-fernet-keys\") pod \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.584083 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gv89r\" (UniqueName: \"kubernetes.io/projected/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-kube-api-access-gv89r\") pod \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.584147 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-combined-ca-bundle\") pod \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.584235 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-config-data\") pod \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\" (UID: \"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264\") " Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.591263 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-kube-api-access-gv89r" (OuterVolumeSpecName: "kube-api-access-gv89r") pod "25286b3f-a5eb-49ce-a0dc-e2a8e99d9264" (UID: "25286b3f-a5eb-49ce-a0dc-e2a8e99d9264"). InnerVolumeSpecName "kube-api-access-gv89r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.591369 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "25286b3f-a5eb-49ce-a0dc-e2a8e99d9264" (UID: "25286b3f-a5eb-49ce-a0dc-e2a8e99d9264"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.645923 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25286b3f-a5eb-49ce-a0dc-e2a8e99d9264" (UID: "25286b3f-a5eb-49ce-a0dc-e2a8e99d9264"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.677989 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-config-data" (OuterVolumeSpecName: "config-data") pod "25286b3f-a5eb-49ce-a0dc-e2a8e99d9264" (UID: "25286b3f-a5eb-49ce-a0dc-e2a8e99d9264"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.686450 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.686484 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.686493 4735 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 15 21:01:06 crc kubenswrapper[4735]: I0215 21:01:06.686507 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gv89r\" (UniqueName: \"kubernetes.io/projected/25286b3f-a5eb-49ce-a0dc-e2a8e99d9264-kube-api-access-gv89r\") on node \"crc\" DevicePath \"\"" Feb 15 21:01:07 crc kubenswrapper[4735]: I0215 21:01:07.199665 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29519821-8vcgr" event={"ID":"25286b3f-a5eb-49ce-a0dc-e2a8e99d9264","Type":"ContainerDied","Data":"5f399ec52e45d02969d0d17bab7f85492a779bfffa96460d80fdda2f9ab76946"} Feb 15 21:01:07 crc kubenswrapper[4735]: I0215 21:01:07.199711 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f399ec52e45d02969d0d17bab7f85492a779bfffa96460d80fdda2f9ab76946" Feb 15 21:01:07 crc kubenswrapper[4735]: I0215 21:01:07.199772 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29519821-8vcgr" Feb 15 21:01:10 crc kubenswrapper[4735]: I0215 21:01:10.490867 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5gxdj" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="registry-server" probeResult="failure" output=< Feb 15 21:01:10 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:01:10 crc kubenswrapper[4735]: > Feb 15 21:01:20 crc kubenswrapper[4735]: I0215 21:01:20.491492 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5gxdj" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="registry-server" probeResult="failure" output=< Feb 15 21:01:20 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:01:20 crc kubenswrapper[4735]: > Feb 15 21:01:29 crc kubenswrapper[4735]: I0215 21:01:29.481429 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:01:29 crc kubenswrapper[4735]: I0215 21:01:29.532463 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:01:29 crc kubenswrapper[4735]: I0215 21:01:29.713849 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5gxdj"] Feb 15 21:01:30 crc kubenswrapper[4735]: I0215 21:01:30.599481 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5gxdj" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="registry-server" containerID="cri-o://022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81" gracePeriod=2 Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.164073 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.178997 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-utilities\") pod \"94047cad-d62c-418d-aef3-a117a51a7ac9\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.179153 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-catalog-content\") pod \"94047cad-d62c-418d-aef3-a117a51a7ac9\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.179250 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwzzd\" (UniqueName: \"kubernetes.io/projected/94047cad-d62c-418d-aef3-a117a51a7ac9-kube-api-access-jwzzd\") pod \"94047cad-d62c-418d-aef3-a117a51a7ac9\" (UID: \"94047cad-d62c-418d-aef3-a117a51a7ac9\") " Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.179849 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-utilities" (OuterVolumeSpecName: "utilities") pod "94047cad-d62c-418d-aef3-a117a51a7ac9" (UID: "94047cad-d62c-418d-aef3-a117a51a7ac9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.186068 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94047cad-d62c-418d-aef3-a117a51a7ac9-kube-api-access-jwzzd" (OuterVolumeSpecName: "kube-api-access-jwzzd") pod "94047cad-d62c-418d-aef3-a117a51a7ac9" (UID: "94047cad-d62c-418d-aef3-a117a51a7ac9"). InnerVolumeSpecName "kube-api-access-jwzzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.280812 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwzzd\" (UniqueName: \"kubernetes.io/projected/94047cad-d62c-418d-aef3-a117a51a7ac9-kube-api-access-jwzzd\") on node \"crc\" DevicePath \"\"" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.280842 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.306293 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94047cad-d62c-418d-aef3-a117a51a7ac9" (UID: "94047cad-d62c-418d-aef3-a117a51a7ac9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.382412 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94047cad-d62c-418d-aef3-a117a51a7ac9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.610358 4735 generic.go:334] "Generic (PLEG): container finished" podID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerID="022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81" exitCode=0 Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.610415 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gxdj" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.610427 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gxdj" event={"ID":"94047cad-d62c-418d-aef3-a117a51a7ac9","Type":"ContainerDied","Data":"022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81"} Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.610606 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gxdj" event={"ID":"94047cad-d62c-418d-aef3-a117a51a7ac9","Type":"ContainerDied","Data":"dcbdcf5171937303a8713654f3eb654ed8e462bbaecddb613c4bbe5012643cf4"} Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.610653 4735 scope.go:117] "RemoveContainer" containerID="022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.652447 4735 scope.go:117] "RemoveContainer" containerID="25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.664085 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5gxdj"] Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.673787 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5gxdj"] Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.700070 4735 scope.go:117] "RemoveContainer" containerID="59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.722570 4735 scope.go:117] "RemoveContainer" containerID="022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81" Feb 15 21:01:31 crc kubenswrapper[4735]: E0215 21:01:31.725187 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81\": container with ID starting with 022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81 not found: ID does not exist" containerID="022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.725218 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81"} err="failed to get container status \"022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81\": rpc error: code = NotFound desc = could not find container \"022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81\": container with ID starting with 022d1d7ce142fcc12ec01713a95dd6cb02d0a63efda3744b316227c0a037ca81 not found: ID does not exist" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.725237 4735 scope.go:117] "RemoveContainer" containerID="25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92" Feb 15 21:01:31 crc kubenswrapper[4735]: E0215 21:01:31.725476 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92\": container with ID starting with 25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92 not found: ID does not exist" containerID="25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.725499 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92"} err="failed to get container status \"25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92\": rpc error: code = NotFound desc = could not find container \"25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92\": container with ID starting with 25d8ea855785b1af4e30f7f3c32c2d57184815ccb45ed88035386654b91c3c92 not found: ID does not exist" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.725511 4735 scope.go:117] "RemoveContainer" containerID="59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9" Feb 15 21:01:31 crc kubenswrapper[4735]: E0215 21:01:31.725782 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9\": container with ID starting with 59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9 not found: ID does not exist" containerID="59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9" Feb 15 21:01:31 crc kubenswrapper[4735]: I0215 21:01:31.725833 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9"} err="failed to get container status \"59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9\": rpc error: code = NotFound desc = could not find container \"59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9\": container with ID starting with 59978f14cab32e25c06cb0338daf3b637f4fcb76263c112f62a689335803bdf9 not found: ID does not exist" Feb 15 21:01:32 crc kubenswrapper[4735]: I0215 21:01:32.923319 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" path="/var/lib/kubelet/pods/94047cad-d62c-418d-aef3-a117a51a7ac9/volumes" Feb 15 21:02:49 crc kubenswrapper[4735]: I0215 21:02:49.679485 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:02:49 crc kubenswrapper[4735]: I0215 21:02:49.680041 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:03:05 crc kubenswrapper[4735]: I0215 21:03:05.560715 4735 generic.go:334] "Generic (PLEG): container finished" podID="25a254b5-abc4-45fa-aba5-935d79ae8959" containerID="1c4da665bc4d39996eff53f8e95fc779c1dce1aba8c1062e0a464a10b8486c32" exitCode=0 Feb 15 21:03:05 crc kubenswrapper[4735]: I0215 21:03:05.560778 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" event={"ID":"25a254b5-abc4-45fa-aba5-935d79ae8959","Type":"ContainerDied","Data":"1c4da665bc4d39996eff53f8e95fc779c1dce1aba8c1062e0a464a10b8486c32"} Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.033140 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.156881 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-0\") pod \"25a254b5-abc4-45fa-aba5-935d79ae8959\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.157049 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-2\") pod \"25a254b5-abc4-45fa-aba5-935d79ae8959\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.157086 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-1\") pod \"25a254b5-abc4-45fa-aba5-935d79ae8959\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.157204 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ssh-key-openstack-edpm-ipam\") pod \"25a254b5-abc4-45fa-aba5-935d79ae8959\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.157246 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-inventory\") pod \"25a254b5-abc4-45fa-aba5-935d79ae8959\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.157274 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-telemetry-combined-ca-bundle\") pod \"25a254b5-abc4-45fa-aba5-935d79ae8959\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.157309 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbq5x\" (UniqueName: \"kubernetes.io/projected/25a254b5-abc4-45fa-aba5-935d79ae8959-kube-api-access-gbq5x\") pod \"25a254b5-abc4-45fa-aba5-935d79ae8959\" (UID: \"25a254b5-abc4-45fa-aba5-935d79ae8959\") " Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.163719 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25a254b5-abc4-45fa-aba5-935d79ae8959-kube-api-access-gbq5x" (OuterVolumeSpecName: "kube-api-access-gbq5x") pod "25a254b5-abc4-45fa-aba5-935d79ae8959" (UID: "25a254b5-abc4-45fa-aba5-935d79ae8959"). InnerVolumeSpecName "kube-api-access-gbq5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.179924 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "25a254b5-abc4-45fa-aba5-935d79ae8959" (UID: "25a254b5-abc4-45fa-aba5-935d79ae8959"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.189864 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-inventory" (OuterVolumeSpecName: "inventory") pod "25a254b5-abc4-45fa-aba5-935d79ae8959" (UID: "25a254b5-abc4-45fa-aba5-935d79ae8959"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.198905 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "25a254b5-abc4-45fa-aba5-935d79ae8959" (UID: "25a254b5-abc4-45fa-aba5-935d79ae8959"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.200495 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "25a254b5-abc4-45fa-aba5-935d79ae8959" (UID: "25a254b5-abc4-45fa-aba5-935d79ae8959"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.201126 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "25a254b5-abc4-45fa-aba5-935d79ae8959" (UID: "25a254b5-abc4-45fa-aba5-935d79ae8959"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.227151 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "25a254b5-abc4-45fa-aba5-935d79ae8959" (UID: "25a254b5-abc4-45fa-aba5-935d79ae8959"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.259143 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.259189 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.259205 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.259222 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-inventory\") on node \"crc\" DevicePath \"\"" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.259237 4735 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.259248 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbq5x\" (UniqueName: \"kubernetes.io/projected/25a254b5-abc4-45fa-aba5-935d79ae8959-kube-api-access-gbq5x\") on node \"crc\" DevicePath \"\"" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.259260 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/25a254b5-abc4-45fa-aba5-935d79ae8959-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.582050 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" event={"ID":"25a254b5-abc4-45fa-aba5-935d79ae8959","Type":"ContainerDied","Data":"0fbe16e878659bc737992b815754348b204c7f56d0c44aa60333a1e34f367ce2"} Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.582431 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fbe16e878659bc737992b815754348b204c7f56d0c44aa60333a1e34f367ce2" Feb 15 21:03:07 crc kubenswrapper[4735]: I0215 21:03:07.582188 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf" Feb 15 21:03:19 crc kubenswrapper[4735]: I0215 21:03:19.680173 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:03:19 crc kubenswrapper[4735]: I0215 21:03:19.680744 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:03:49 crc kubenswrapper[4735]: I0215 21:03:49.680263 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:03:49 crc kubenswrapper[4735]: I0215 21:03:49.682146 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:03:49 crc kubenswrapper[4735]: I0215 21:03:49.682249 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 21:03:49 crc kubenswrapper[4735]: I0215 21:03:49.683252 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ad3865af6e8d595faa44438b03b1d22c4d0d2868c6797e3edfc5272ac148174c"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 21:03:49 crc kubenswrapper[4735]: I0215 21:03:49.683376 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://ad3865af6e8d595faa44438b03b1d22c4d0d2868c6797e3edfc5272ac148174c" gracePeriod=600 Feb 15 21:03:50 crc kubenswrapper[4735]: I0215 21:03:50.096802 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="ad3865af6e8d595faa44438b03b1d22c4d0d2868c6797e3edfc5272ac148174c" exitCode=0 Feb 15 21:03:50 crc kubenswrapper[4735]: I0215 21:03:50.096882 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"ad3865af6e8d595faa44438b03b1d22c4d0d2868c6797e3edfc5272ac148174c"} Feb 15 21:03:50 crc kubenswrapper[4735]: I0215 21:03:50.097102 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4"} Feb 15 21:03:50 crc kubenswrapper[4735]: I0215 21:03:50.097120 4735 scope.go:117] "RemoveContainer" containerID="f242d5868b8ea52c827e375a3985c3dcf5172a520c1bd13b3c1ece95ea185639" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.258624 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c5fjx"] Feb 15 21:03:52 crc kubenswrapper[4735]: E0215 21:03:52.259563 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25a254b5-abc4-45fa-aba5-935d79ae8959" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.259584 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="25a254b5-abc4-45fa-aba5-935d79ae8959" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 15 21:03:52 crc kubenswrapper[4735]: E0215 21:03:52.259607 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25286b3f-a5eb-49ce-a0dc-e2a8e99d9264" containerName="keystone-cron" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.259615 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="25286b3f-a5eb-49ce-a0dc-e2a8e99d9264" containerName="keystone-cron" Feb 15 21:03:52 crc kubenswrapper[4735]: E0215 21:03:52.259636 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="extract-utilities" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.259643 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="extract-utilities" Feb 15 21:03:52 crc kubenswrapper[4735]: E0215 21:03:52.259660 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="registry-server" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.259670 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="registry-server" Feb 15 21:03:52 crc kubenswrapper[4735]: E0215 21:03:52.259689 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="extract-content" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.259695 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="extract-content" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.259911 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="94047cad-d62c-418d-aef3-a117a51a7ac9" containerName="registry-server" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.259923 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="25a254b5-abc4-45fa-aba5-935d79ae8959" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.259934 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="25286b3f-a5eb-49ce-a0dc-e2a8e99d9264" containerName="keystone-cron" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.261525 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.285904 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c5fjx"] Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.300052 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-catalog-content\") pod \"community-operators-c5fjx\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.300111 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5tz5\" (UniqueName: \"kubernetes.io/projected/3078a475-3cd1-493b-8279-963f1a8fdcd3-kube-api-access-x5tz5\") pod \"community-operators-c5fjx\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.300241 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-utilities\") pod \"community-operators-c5fjx\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.401728 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-utilities\") pod \"community-operators-c5fjx\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.402120 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-catalog-content\") pod \"community-operators-c5fjx\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.402321 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5tz5\" (UniqueName: \"kubernetes.io/projected/3078a475-3cd1-493b-8279-963f1a8fdcd3-kube-api-access-x5tz5\") pod \"community-operators-c5fjx\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.402194 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-utilities\") pod \"community-operators-c5fjx\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.403196 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-catalog-content\") pod \"community-operators-c5fjx\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.434489 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5tz5\" (UniqueName: \"kubernetes.io/projected/3078a475-3cd1-493b-8279-963f1a8fdcd3-kube-api-access-x5tz5\") pod \"community-operators-c5fjx\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:52 crc kubenswrapper[4735]: I0215 21:03:52.578528 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:03:53 crc kubenswrapper[4735]: I0215 21:03:53.179319 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c5fjx"] Feb 15 21:03:53 crc kubenswrapper[4735]: W0215 21:03:53.189506 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3078a475_3cd1_493b_8279_963f1a8fdcd3.slice/crio-6cf7e9aa6a430dbeb1b270c949da0c6f2ec5c8e4d1a780253d58da57c2f18fa8 WatchSource:0}: Error finding container 6cf7e9aa6a430dbeb1b270c949da0c6f2ec5c8e4d1a780253d58da57c2f18fa8: Status 404 returned error can't find the container with id 6cf7e9aa6a430dbeb1b270c949da0c6f2ec5c8e4d1a780253d58da57c2f18fa8 Feb 15 21:03:54 crc kubenswrapper[4735]: I0215 21:03:54.153799 4735 generic.go:334] "Generic (PLEG): container finished" podID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerID="6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e" exitCode=0 Feb 15 21:03:54 crc kubenswrapper[4735]: I0215 21:03:54.154010 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5fjx" event={"ID":"3078a475-3cd1-493b-8279-963f1a8fdcd3","Type":"ContainerDied","Data":"6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e"} Feb 15 21:03:54 crc kubenswrapper[4735]: I0215 21:03:54.154066 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5fjx" event={"ID":"3078a475-3cd1-493b-8279-963f1a8fdcd3","Type":"ContainerStarted","Data":"6cf7e9aa6a430dbeb1b270c949da0c6f2ec5c8e4d1a780253d58da57c2f18fa8"} Feb 15 21:03:55 crc kubenswrapper[4735]: I0215 21:03:55.163449 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5fjx" event={"ID":"3078a475-3cd1-493b-8279-963f1a8fdcd3","Type":"ContainerStarted","Data":"2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf"} Feb 15 21:03:57 crc kubenswrapper[4735]: I0215 21:03:57.190134 4735 generic.go:334] "Generic (PLEG): container finished" podID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerID="2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf" exitCode=0 Feb 15 21:03:57 crc kubenswrapper[4735]: I0215 21:03:57.190188 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5fjx" event={"ID":"3078a475-3cd1-493b-8279-963f1a8fdcd3","Type":"ContainerDied","Data":"2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf"} Feb 15 21:03:58 crc kubenswrapper[4735]: I0215 21:03:58.200533 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5fjx" event={"ID":"3078a475-3cd1-493b-8279-963f1a8fdcd3","Type":"ContainerStarted","Data":"6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa"} Feb 15 21:03:58 crc kubenswrapper[4735]: I0215 21:03:58.225789 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c5fjx" podStartSLOduration=2.663533065 podStartE2EDuration="6.225770642s" podCreationTimestamp="2026-02-15 21:03:52 +0000 UTC" firstStartedPulling="2026-02-15 21:03:54.155830978 +0000 UTC m=+2842.021846601" lastFinishedPulling="2026-02-15 21:03:57.718068535 +0000 UTC m=+2845.584084178" observedRunningTime="2026-02-15 21:03:58.218547797 +0000 UTC m=+2846.084563420" watchObservedRunningTime="2026-02-15 21:03:58.225770642 +0000 UTC m=+2846.091786265" Feb 15 21:04:02 crc kubenswrapper[4735]: I0215 21:04:02.579415 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:04:02 crc kubenswrapper[4735]: I0215 21:04:02.580259 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:04:02 crc kubenswrapper[4735]: I0215 21:04:02.650825 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:04:03 crc kubenswrapper[4735]: I0215 21:04:03.301341 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:04:03 crc kubenswrapper[4735]: I0215 21:04:03.345974 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c5fjx"] Feb 15 21:04:05 crc kubenswrapper[4735]: I0215 21:04:05.306355 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c5fjx" podUID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerName="registry-server" containerID="cri-o://6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa" gracePeriod=2 Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.274231 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.325868 4735 generic.go:334] "Generic (PLEG): container finished" podID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerID="6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa" exitCode=0 Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.325909 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5fjx" event={"ID":"3078a475-3cd1-493b-8279-963f1a8fdcd3","Type":"ContainerDied","Data":"6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa"} Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.325954 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5fjx" event={"ID":"3078a475-3cd1-493b-8279-963f1a8fdcd3","Type":"ContainerDied","Data":"6cf7e9aa6a430dbeb1b270c949da0c6f2ec5c8e4d1a780253d58da57c2f18fa8"} Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.325975 4735 scope.go:117] "RemoveContainer" containerID="6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.326096 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5fjx" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.346081 4735 scope.go:117] "RemoveContainer" containerID="2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.366549 4735 scope.go:117] "RemoveContainer" containerID="6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.371690 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-utilities\") pod \"3078a475-3cd1-493b-8279-963f1a8fdcd3\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.371830 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-catalog-content\") pod \"3078a475-3cd1-493b-8279-963f1a8fdcd3\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.371887 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5tz5\" (UniqueName: \"kubernetes.io/projected/3078a475-3cd1-493b-8279-963f1a8fdcd3-kube-api-access-x5tz5\") pod \"3078a475-3cd1-493b-8279-963f1a8fdcd3\" (UID: \"3078a475-3cd1-493b-8279-963f1a8fdcd3\") " Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.372585 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-utilities" (OuterVolumeSpecName: "utilities") pod "3078a475-3cd1-493b-8279-963f1a8fdcd3" (UID: "3078a475-3cd1-493b-8279-963f1a8fdcd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.378963 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3078a475-3cd1-493b-8279-963f1a8fdcd3-kube-api-access-x5tz5" (OuterVolumeSpecName: "kube-api-access-x5tz5") pod "3078a475-3cd1-493b-8279-963f1a8fdcd3" (UID: "3078a475-3cd1-493b-8279-963f1a8fdcd3"). InnerVolumeSpecName "kube-api-access-x5tz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.430966 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3078a475-3cd1-493b-8279-963f1a8fdcd3" (UID: "3078a475-3cd1-493b-8279-963f1a8fdcd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.454693 4735 scope.go:117] "RemoveContainer" containerID="6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa" Feb 15 21:04:06 crc kubenswrapper[4735]: E0215 21:04:06.455170 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa\": container with ID starting with 6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa not found: ID does not exist" containerID="6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.455219 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa"} err="failed to get container status \"6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa\": rpc error: code = NotFound desc = could not find container \"6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa\": container with ID starting with 6ffcb58b11a120c406def9dcb465b2ccd853402ee429aedeb0eedbb0d8bcb7aa not found: ID does not exist" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.455328 4735 scope.go:117] "RemoveContainer" containerID="2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf" Feb 15 21:04:06 crc kubenswrapper[4735]: E0215 21:04:06.455739 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf\": container with ID starting with 2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf not found: ID does not exist" containerID="2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.455771 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf"} err="failed to get container status \"2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf\": rpc error: code = NotFound desc = could not find container \"2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf\": container with ID starting with 2381c34aff3bed677a514a297e0577e49ebbfef04c4e726e58cd5214b413febf not found: ID does not exist" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.455796 4735 scope.go:117] "RemoveContainer" containerID="6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e" Feb 15 21:04:06 crc kubenswrapper[4735]: E0215 21:04:06.456054 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e\": container with ID starting with 6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e not found: ID does not exist" containerID="6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.456079 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e"} err="failed to get container status \"6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e\": rpc error: code = NotFound desc = could not find container \"6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e\": container with ID starting with 6696d793d14de667193f9d21d4abf48c27b3ae1c4960e7c0bd20183e1410461e not found: ID does not exist" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.474331 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.474439 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5tz5\" (UniqueName: \"kubernetes.io/projected/3078a475-3cd1-493b-8279-963f1a8fdcd3-kube-api-access-x5tz5\") on node \"crc\" DevicePath \"\"" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.474499 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3078a475-3cd1-493b-8279-963f1a8fdcd3-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.655653 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c5fjx"] Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.664540 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c5fjx"] Feb 15 21:04:06 crc kubenswrapper[4735]: I0215 21:04:06.907846 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3078a475-3cd1-493b-8279-963f1a8fdcd3" path="/var/lib/kubelet/pods/3078a475-3cd1-493b-8279-963f1a8fdcd3/volumes" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.166496 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 15 21:04:08 crc kubenswrapper[4735]: E0215 21:04:08.172704 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerName="extract-content" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.172764 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerName="extract-content" Feb 15 21:04:08 crc kubenswrapper[4735]: E0215 21:04:08.172837 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerName="extract-utilities" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.172853 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerName="extract-utilities" Feb 15 21:04:08 crc kubenswrapper[4735]: E0215 21:04:08.172886 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerName="registry-server" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.172899 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerName="registry-server" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.173244 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3078a475-3cd1-493b-8279-963f1a8fdcd3" containerName="registry-server" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.174238 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.179500 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.179522 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.179651 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.179756 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-j75vs" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.188872 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.347085 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.347178 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.347226 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.347251 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psbch\" (UniqueName: \"kubernetes.io/projected/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-kube-api-access-psbch\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.347271 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.347294 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.347340 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.347366 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.347434 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-config-data\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.448956 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-config-data\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.449293 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.449447 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.449566 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.449667 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psbch\" (UniqueName: \"kubernetes.io/projected/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-kube-api-access-psbch\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.449762 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.449866 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.450009 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.450492 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.450501 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.450025 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-config-data\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.451005 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.451125 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.451328 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.458801 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.458916 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.462801 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.466726 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psbch\" (UniqueName: \"kubernetes.io/projected/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-kube-api-access-psbch\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.482877 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " pod="openstack/tempest-tests-tempest" Feb 15 21:04:08 crc kubenswrapper[4735]: I0215 21:04:08.528583 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 15 21:04:09 crc kubenswrapper[4735]: I0215 21:04:09.027005 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 21:04:09 crc kubenswrapper[4735]: I0215 21:04:09.031460 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 15 21:04:09 crc kubenswrapper[4735]: I0215 21:04:09.349963 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383","Type":"ContainerStarted","Data":"5a4e6b489f4c322458d9ff0b86f8a712a576a93e47188959ad07a2cb7f80f137"} Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:16.470757 4735 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-nhpdn container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:16.470812 4735 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-nhpdn container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:16.471258 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" podUID="9ec3f4d1-68c8-4460-8802-bc4ac3c1014f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:16.471355 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nhpdn" podUID="9ec3f4d1-68c8-4460-8802-bc4ac3c1014f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:17.006949 4735 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qmtqd container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:17.007005 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" podUID="27821f2f-df7e-42d2-b687-244c7b6e5a3f" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:17.007013 4735 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qmtqd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:17.007040 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qmtqd" podUID="27821f2f-df7e-42d2-b687-244c7b6e5a3f" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:17.033955 4735 patch_prober.go:28] interesting pod/console-55bc4b8956-6szxr container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:17.034004 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-55bc4b8956-6szxr" podUID="f3aed0a9-28fa-4d6a-a8ca-1d657e5983d0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.35:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:17.095124 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:17.095189 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 15 21:04:17 crc kubenswrapper[4735]: I0215 21:04:17.797291 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-lkzd9" podUID="fc8fe06e-5292-4b4c-9b20-b2a3f4fce798" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 15 21:04:35 crc kubenswrapper[4735]: I0215 21:04:35.888486 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zpl8t"] Feb 15 21:04:35 crc kubenswrapper[4735]: I0215 21:04:35.892930 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:35 crc kubenswrapper[4735]: I0215 21:04:35.914971 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpl8t"] Feb 15 21:04:35 crc kubenswrapper[4735]: I0215 21:04:35.989354 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-catalog-content\") pod \"redhat-marketplace-zpl8t\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:35 crc kubenswrapper[4735]: I0215 21:04:35.989579 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-utilities\") pod \"redhat-marketplace-zpl8t\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:35 crc kubenswrapper[4735]: I0215 21:04:35.989616 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd6b4\" (UniqueName: \"kubernetes.io/projected/45c7958f-09f8-48ec-9008-4be897617cfd-kube-api-access-zd6b4\") pod \"redhat-marketplace-zpl8t\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:36 crc kubenswrapper[4735]: I0215 21:04:36.091029 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-utilities\") pod \"redhat-marketplace-zpl8t\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:36 crc kubenswrapper[4735]: I0215 21:04:36.091136 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd6b4\" (UniqueName: \"kubernetes.io/projected/45c7958f-09f8-48ec-9008-4be897617cfd-kube-api-access-zd6b4\") pod \"redhat-marketplace-zpl8t\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:36 crc kubenswrapper[4735]: I0215 21:04:36.091204 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-catalog-content\") pod \"redhat-marketplace-zpl8t\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:36 crc kubenswrapper[4735]: I0215 21:04:36.091917 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-catalog-content\") pod \"redhat-marketplace-zpl8t\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:36 crc kubenswrapper[4735]: I0215 21:04:36.092521 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-utilities\") pod \"redhat-marketplace-zpl8t\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:36 crc kubenswrapper[4735]: I0215 21:04:36.184309 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd6b4\" (UniqueName: \"kubernetes.io/projected/45c7958f-09f8-48ec-9008-4be897617cfd-kube-api-access-zd6b4\") pod \"redhat-marketplace-zpl8t\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:36 crc kubenswrapper[4735]: I0215 21:04:36.230179 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:04:52 crc kubenswrapper[4735]: E0215 21:04:52.799957 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 15 21:04:52 crc kubenswrapper[4735]: E0215 21:04:52.818561 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-psbch,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(abcf2b9d-8410-4a1b-be0e-3dde2a6e3383): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 15 21:04:52 crc kubenswrapper[4735]: E0215 21:04:52.822016 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" Feb 15 21:04:53 crc kubenswrapper[4735]: I0215 21:04:53.444361 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpl8t"] Feb 15 21:04:53 crc kubenswrapper[4735]: I0215 21:04:53.766044 4735 generic.go:334] "Generic (PLEG): container finished" podID="45c7958f-09f8-48ec-9008-4be897617cfd" containerID="fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd" exitCode=0 Feb 15 21:04:53 crc kubenswrapper[4735]: I0215 21:04:53.766100 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpl8t" event={"ID":"45c7958f-09f8-48ec-9008-4be897617cfd","Type":"ContainerDied","Data":"fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd"} Feb 15 21:04:53 crc kubenswrapper[4735]: I0215 21:04:53.766151 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpl8t" event={"ID":"45c7958f-09f8-48ec-9008-4be897617cfd","Type":"ContainerStarted","Data":"5aa55a6fdc7aad433305006fc71226891dc5ddc179b1e62d5985522e6370be58"} Feb 15 21:04:53 crc kubenswrapper[4735]: E0215 21:04:53.768190 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" Feb 15 21:04:55 crc kubenswrapper[4735]: I0215 21:04:55.794273 4735 generic.go:334] "Generic (PLEG): container finished" podID="45c7958f-09f8-48ec-9008-4be897617cfd" containerID="716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df" exitCode=0 Feb 15 21:04:55 crc kubenswrapper[4735]: I0215 21:04:55.794339 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpl8t" event={"ID":"45c7958f-09f8-48ec-9008-4be897617cfd","Type":"ContainerDied","Data":"716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df"} Feb 15 21:04:56 crc kubenswrapper[4735]: I0215 21:04:56.812156 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpl8t" event={"ID":"45c7958f-09f8-48ec-9008-4be897617cfd","Type":"ContainerStarted","Data":"26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2"} Feb 15 21:05:04 crc kubenswrapper[4735]: I0215 21:05:04.914522 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zpl8t" podStartSLOduration=27.495338493 podStartE2EDuration="29.914501476s" podCreationTimestamp="2026-02-15 21:04:35 +0000 UTC" firstStartedPulling="2026-02-15 21:04:53.767935864 +0000 UTC m=+2901.633951487" lastFinishedPulling="2026-02-15 21:04:56.187098847 +0000 UTC m=+2904.053114470" observedRunningTime="2026-02-15 21:04:56.848880751 +0000 UTC m=+2904.714896404" watchObservedRunningTime="2026-02-15 21:05:04.914501476 +0000 UTC m=+2912.780517099" Feb 15 21:05:05 crc kubenswrapper[4735]: I0215 21:05:05.458318 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 15 21:05:06 crc kubenswrapper[4735]: I0215 21:05:06.231231 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:05:06 crc kubenswrapper[4735]: I0215 21:05:06.231412 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:05:06 crc kubenswrapper[4735]: I0215 21:05:06.288656 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:05:06 crc kubenswrapper[4735]: I0215 21:05:06.971489 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:05:07 crc kubenswrapper[4735]: I0215 21:05:07.074677 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpl8t"] Feb 15 21:05:07 crc kubenswrapper[4735]: I0215 21:05:07.934105 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383","Type":"ContainerStarted","Data":"89280b91de1b8c16643f8e09110b4c71cba2e2e98b4c24d80ae67cc5a38fde2a"} Feb 15 21:05:07 crc kubenswrapper[4735]: I0215 21:05:07.951878 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.523298319 podStartE2EDuration="1m0.951860723s" podCreationTimestamp="2026-02-15 21:04:07 +0000 UTC" firstStartedPulling="2026-02-15 21:04:09.026763538 +0000 UTC m=+2856.892779151" lastFinishedPulling="2026-02-15 21:05:05.455325922 +0000 UTC m=+2913.321341555" observedRunningTime="2026-02-15 21:05:07.947449054 +0000 UTC m=+2915.813464677" watchObservedRunningTime="2026-02-15 21:05:07.951860723 +0000 UTC m=+2915.817876346" Feb 15 21:05:08 crc kubenswrapper[4735]: I0215 21:05:08.941899 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zpl8t" podUID="45c7958f-09f8-48ec-9008-4be897617cfd" containerName="registry-server" containerID="cri-o://26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2" gracePeriod=2 Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.402083 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.507286 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd6b4\" (UniqueName: \"kubernetes.io/projected/45c7958f-09f8-48ec-9008-4be897617cfd-kube-api-access-zd6b4\") pod \"45c7958f-09f8-48ec-9008-4be897617cfd\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.507511 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-catalog-content\") pod \"45c7958f-09f8-48ec-9008-4be897617cfd\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.507558 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-utilities\") pod \"45c7958f-09f8-48ec-9008-4be897617cfd\" (UID: \"45c7958f-09f8-48ec-9008-4be897617cfd\") " Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.509526 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-utilities" (OuterVolumeSpecName: "utilities") pod "45c7958f-09f8-48ec-9008-4be897617cfd" (UID: "45c7958f-09f8-48ec-9008-4be897617cfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.532826 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45c7958f-09f8-48ec-9008-4be897617cfd-kube-api-access-zd6b4" (OuterVolumeSpecName: "kube-api-access-zd6b4") pod "45c7958f-09f8-48ec-9008-4be897617cfd" (UID: "45c7958f-09f8-48ec-9008-4be897617cfd"). InnerVolumeSpecName "kube-api-access-zd6b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.535899 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45c7958f-09f8-48ec-9008-4be897617cfd" (UID: "45c7958f-09f8-48ec-9008-4be897617cfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.610287 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.610327 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c7958f-09f8-48ec-9008-4be897617cfd-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.610341 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd6b4\" (UniqueName: \"kubernetes.io/projected/45c7958f-09f8-48ec-9008-4be897617cfd-kube-api-access-zd6b4\") on node \"crc\" DevicePath \"\"" Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.951562 4735 generic.go:334] "Generic (PLEG): container finished" podID="45c7958f-09f8-48ec-9008-4be897617cfd" containerID="26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2" exitCode=0 Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.951620 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zpl8t" Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.951642 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpl8t" event={"ID":"45c7958f-09f8-48ec-9008-4be897617cfd","Type":"ContainerDied","Data":"26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2"} Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.952300 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpl8t" event={"ID":"45c7958f-09f8-48ec-9008-4be897617cfd","Type":"ContainerDied","Data":"5aa55a6fdc7aad433305006fc71226891dc5ddc179b1e62d5985522e6370be58"} Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.952337 4735 scope.go:117] "RemoveContainer" containerID="26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2" Feb 15 21:05:09 crc kubenswrapper[4735]: I0215 21:05:09.983058 4735 scope.go:117] "RemoveContainer" containerID="716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df" Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.010881 4735 scope.go:117] "RemoveContainer" containerID="fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd" Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.037118 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpl8t"] Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.045296 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpl8t"] Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.052983 4735 scope.go:117] "RemoveContainer" containerID="26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2" Feb 15 21:05:10 crc kubenswrapper[4735]: E0215 21:05:10.053381 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2\": container with ID starting with 26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2 not found: ID does not exist" containerID="26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2" Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.053434 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2"} err="failed to get container status \"26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2\": rpc error: code = NotFound desc = could not find container \"26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2\": container with ID starting with 26029884fe73b039150aac6ecf16f80c48c2489e8f3f8770c43ba4746a9e34a2 not found: ID does not exist" Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.053469 4735 scope.go:117] "RemoveContainer" containerID="716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df" Feb 15 21:05:10 crc kubenswrapper[4735]: E0215 21:05:10.053745 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df\": container with ID starting with 716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df not found: ID does not exist" containerID="716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df" Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.053776 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df"} err="failed to get container status \"716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df\": rpc error: code = NotFound desc = could not find container \"716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df\": container with ID starting with 716119887e644677461d849ae3803db7d4fdb8f95606c9bd1a25f83a97a7e3df not found: ID does not exist" Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.053796 4735 scope.go:117] "RemoveContainer" containerID="fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd" Feb 15 21:05:10 crc kubenswrapper[4735]: E0215 21:05:10.054210 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd\": container with ID starting with fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd not found: ID does not exist" containerID="fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd" Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.054241 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd"} err="failed to get container status \"fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd\": rpc error: code = NotFound desc = could not find container \"fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd\": container with ID starting with fe78a63361d41b552361df0a006102d3ea120557099b2bf43eb3ee3d8ee570fd not found: ID does not exist" Feb 15 21:05:10 crc kubenswrapper[4735]: I0215 21:05:10.907240 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45c7958f-09f8-48ec-9008-4be897617cfd" path="/var/lib/kubelet/pods/45c7958f-09f8-48ec-9008-4be897617cfd/volumes" Feb 15 21:05:49 crc kubenswrapper[4735]: I0215 21:05:49.679559 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:05:49 crc kubenswrapper[4735]: I0215 21:05:49.680109 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:06:19 crc kubenswrapper[4735]: I0215 21:06:19.679985 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:06:19 crc kubenswrapper[4735]: I0215 21:06:19.680497 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:06:49 crc kubenswrapper[4735]: I0215 21:06:49.679737 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:06:49 crc kubenswrapper[4735]: I0215 21:06:49.680463 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:06:49 crc kubenswrapper[4735]: I0215 21:06:49.680528 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 21:06:49 crc kubenswrapper[4735]: I0215 21:06:49.681563 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 21:06:49 crc kubenswrapper[4735]: I0215 21:06:49.681629 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" gracePeriod=600 Feb 15 21:06:49 crc kubenswrapper[4735]: E0215 21:06:49.826883 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:06:50 crc kubenswrapper[4735]: I0215 21:06:50.820192 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" exitCode=0 Feb 15 21:06:50 crc kubenswrapper[4735]: I0215 21:06:50.820224 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4"} Feb 15 21:06:50 crc kubenswrapper[4735]: I0215 21:06:50.820268 4735 scope.go:117] "RemoveContainer" containerID="ad3865af6e8d595faa44438b03b1d22c4d0d2868c6797e3edfc5272ac148174c" Feb 15 21:06:50 crc kubenswrapper[4735]: I0215 21:06:50.820768 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:06:50 crc kubenswrapper[4735]: E0215 21:06:50.821179 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.371505 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-96ppb"] Feb 15 21:07:01 crc kubenswrapper[4735]: E0215 21:07:01.372509 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c7958f-09f8-48ec-9008-4be897617cfd" containerName="extract-content" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.372530 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c7958f-09f8-48ec-9008-4be897617cfd" containerName="extract-content" Feb 15 21:07:01 crc kubenswrapper[4735]: E0215 21:07:01.372545 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c7958f-09f8-48ec-9008-4be897617cfd" containerName="extract-utilities" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.372551 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c7958f-09f8-48ec-9008-4be897617cfd" containerName="extract-utilities" Feb 15 21:07:01 crc kubenswrapper[4735]: E0215 21:07:01.372570 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c7958f-09f8-48ec-9008-4be897617cfd" containerName="registry-server" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.372577 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c7958f-09f8-48ec-9008-4be897617cfd" containerName="registry-server" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.372764 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c7958f-09f8-48ec-9008-4be897617cfd" containerName="registry-server" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.375715 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.463240 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-96ppb"] Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.532098 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-catalog-content\") pod \"certified-operators-96ppb\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.532197 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-utilities\") pod \"certified-operators-96ppb\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.533383 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6vc7\" (UniqueName: \"kubernetes.io/projected/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-kube-api-access-t6vc7\") pod \"certified-operators-96ppb\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.636309 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-utilities\") pod \"certified-operators-96ppb\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.636391 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6vc7\" (UniqueName: \"kubernetes.io/projected/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-kube-api-access-t6vc7\") pod \"certified-operators-96ppb\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.636455 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-catalog-content\") pod \"certified-operators-96ppb\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.636928 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-catalog-content\") pod \"certified-operators-96ppb\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.637052 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-utilities\") pod \"certified-operators-96ppb\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.656081 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6vc7\" (UniqueName: \"kubernetes.io/projected/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-kube-api-access-t6vc7\") pod \"certified-operators-96ppb\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.697717 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:01 crc kubenswrapper[4735]: I0215 21:07:01.886904 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:07:01 crc kubenswrapper[4735]: E0215 21:07:01.887167 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:07:02 crc kubenswrapper[4735]: I0215 21:07:02.700688 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-96ppb"] Feb 15 21:07:02 crc kubenswrapper[4735]: I0215 21:07:02.938222 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-96ppb" event={"ID":"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c","Type":"ContainerStarted","Data":"b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692"} Feb 15 21:07:02 crc kubenswrapper[4735]: I0215 21:07:02.938279 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-96ppb" event={"ID":"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c","Type":"ContainerStarted","Data":"d5e387dc0c943b2e7bb33c4322640c743450d96aead4f97959618342c1797deb"} Feb 15 21:07:03 crc kubenswrapper[4735]: I0215 21:07:03.956369 4735 generic.go:334] "Generic (PLEG): container finished" podID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerID="b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692" exitCode=0 Feb 15 21:07:03 crc kubenswrapper[4735]: I0215 21:07:03.956449 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-96ppb" event={"ID":"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c","Type":"ContainerDied","Data":"b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692"} Feb 15 21:07:05 crc kubenswrapper[4735]: I0215 21:07:05.977418 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-96ppb" event={"ID":"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c","Type":"ContainerStarted","Data":"e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e"} Feb 15 21:07:06 crc kubenswrapper[4735]: I0215 21:07:06.988560 4735 generic.go:334] "Generic (PLEG): container finished" podID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerID="e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e" exitCode=0 Feb 15 21:07:06 crc kubenswrapper[4735]: I0215 21:07:06.988658 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-96ppb" event={"ID":"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c","Type":"ContainerDied","Data":"e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e"} Feb 15 21:07:08 crc kubenswrapper[4735]: I0215 21:07:08.000823 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-96ppb" event={"ID":"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c","Type":"ContainerStarted","Data":"986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215"} Feb 15 21:07:08 crc kubenswrapper[4735]: I0215 21:07:08.033106 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-96ppb" podStartSLOduration=3.6105242669999997 podStartE2EDuration="7.033081315s" podCreationTimestamp="2026-02-15 21:07:01 +0000 UTC" firstStartedPulling="2026-02-15 21:07:03.95866921 +0000 UTC m=+3031.824684833" lastFinishedPulling="2026-02-15 21:07:07.381226258 +0000 UTC m=+3035.247241881" observedRunningTime="2026-02-15 21:07:08.029933461 +0000 UTC m=+3035.895949094" watchObservedRunningTime="2026-02-15 21:07:08.033081315 +0000 UTC m=+3035.899096968" Feb 15 21:07:11 crc kubenswrapper[4735]: I0215 21:07:11.698986 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:11 crc kubenswrapper[4735]: I0215 21:07:11.699526 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:12 crc kubenswrapper[4735]: I0215 21:07:12.744856 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-96ppb" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerName="registry-server" probeResult="failure" output=< Feb 15 21:07:12 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:07:12 crc kubenswrapper[4735]: > Feb 15 21:07:14 crc kubenswrapper[4735]: I0215 21:07:14.887430 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:07:14 crc kubenswrapper[4735]: E0215 21:07:14.888124 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:07:21 crc kubenswrapper[4735]: I0215 21:07:21.746509 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:21 crc kubenswrapper[4735]: I0215 21:07:21.806422 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:21 crc kubenswrapper[4735]: I0215 21:07:21.993738 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-96ppb"] Feb 15 21:07:23 crc kubenswrapper[4735]: I0215 21:07:23.155889 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-96ppb" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerName="registry-server" containerID="cri-o://986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215" gracePeriod=2 Feb 15 21:07:23 crc kubenswrapper[4735]: I0215 21:07:23.844017 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:23 crc kubenswrapper[4735]: I0215 21:07:23.955608 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6vc7\" (UniqueName: \"kubernetes.io/projected/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-kube-api-access-t6vc7\") pod \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " Feb 15 21:07:23 crc kubenswrapper[4735]: I0215 21:07:23.955822 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-catalog-content\") pod \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " Feb 15 21:07:23 crc kubenswrapper[4735]: I0215 21:07:23.955891 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-utilities\") pod \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\" (UID: \"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c\") " Feb 15 21:07:23 crc kubenswrapper[4735]: I0215 21:07:23.956600 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-utilities" (OuterVolumeSpecName: "utilities") pod "5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" (UID: "5fae6e75-eaf2-44c4-b8f1-fa56c642a54c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:07:23 crc kubenswrapper[4735]: I0215 21:07:23.993634 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-kube-api-access-t6vc7" (OuterVolumeSpecName: "kube-api-access-t6vc7") pod "5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" (UID: "5fae6e75-eaf2-44c4-b8f1-fa56c642a54c"). InnerVolumeSpecName "kube-api-access-t6vc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.038881 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" (UID: "5fae6e75-eaf2-44c4-b8f1-fa56c642a54c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.057939 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.057984 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.057996 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6vc7\" (UniqueName: \"kubernetes.io/projected/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c-kube-api-access-t6vc7\") on node \"crc\" DevicePath \"\"" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.168214 4735 generic.go:334] "Generic (PLEG): container finished" podID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerID="986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215" exitCode=0 Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.168254 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-96ppb" event={"ID":"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c","Type":"ContainerDied","Data":"986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215"} Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.168279 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-96ppb" event={"ID":"5fae6e75-eaf2-44c4-b8f1-fa56c642a54c","Type":"ContainerDied","Data":"d5e387dc0c943b2e7bb33c4322640c743450d96aead4f97959618342c1797deb"} Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.168283 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-96ppb" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.168295 4735 scope.go:117] "RemoveContainer" containerID="986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.201938 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-96ppb"] Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.209332 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-96ppb"] Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.220120 4735 scope.go:117] "RemoveContainer" containerID="e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.247295 4735 scope.go:117] "RemoveContainer" containerID="b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.303908 4735 scope.go:117] "RemoveContainer" containerID="986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215" Feb 15 21:07:24 crc kubenswrapper[4735]: E0215 21:07:24.304350 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215\": container with ID starting with 986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215 not found: ID does not exist" containerID="986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.304381 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215"} err="failed to get container status \"986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215\": rpc error: code = NotFound desc = could not find container \"986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215\": container with ID starting with 986ffedde171de8734a8fac2d34ac140359811e61749984c9eb1ebb24f575215 not found: ID does not exist" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.304403 4735 scope.go:117] "RemoveContainer" containerID="e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e" Feb 15 21:07:24 crc kubenswrapper[4735]: E0215 21:07:24.304671 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e\": container with ID starting with e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e not found: ID does not exist" containerID="e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.304699 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e"} err="failed to get container status \"e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e\": rpc error: code = NotFound desc = could not find container \"e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e\": container with ID starting with e33a483c57c78df7d1ef577b17b12a9f77dad16978f01dde9f824befba3ad54e not found: ID does not exist" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.304714 4735 scope.go:117] "RemoveContainer" containerID="b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692" Feb 15 21:07:24 crc kubenswrapper[4735]: E0215 21:07:24.305980 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692\": container with ID starting with b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692 not found: ID does not exist" containerID="b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.306004 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692"} err="failed to get container status \"b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692\": rpc error: code = NotFound desc = could not find container \"b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692\": container with ID starting with b9f9bc3dc5bca690fbc591db511114921b67ec7957116b7598d79d301d3de692 not found: ID does not exist" Feb 15 21:07:24 crc kubenswrapper[4735]: I0215 21:07:24.896762 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" path="/var/lib/kubelet/pods/5fae6e75-eaf2-44c4-b8f1-fa56c642a54c/volumes" Feb 15 21:07:25 crc kubenswrapper[4735]: I0215 21:07:25.886425 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:07:25 crc kubenswrapper[4735]: E0215 21:07:25.887378 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:07:36 crc kubenswrapper[4735]: I0215 21:07:36.887498 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:07:36 crc kubenswrapper[4735]: E0215 21:07:36.888555 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:07:47 crc kubenswrapper[4735]: I0215 21:07:47.886681 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:07:47 crc kubenswrapper[4735]: E0215 21:07:47.887461 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:07:59 crc kubenswrapper[4735]: I0215 21:07:59.887427 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:07:59 crc kubenswrapper[4735]: E0215 21:07:59.888310 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:08:12 crc kubenswrapper[4735]: I0215 21:08:12.902989 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:08:12 crc kubenswrapper[4735]: E0215 21:08:12.905939 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:08:27 crc kubenswrapper[4735]: I0215 21:08:27.886975 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:08:27 crc kubenswrapper[4735]: E0215 21:08:27.888563 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:08:42 crc kubenswrapper[4735]: I0215 21:08:42.894026 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:08:42 crc kubenswrapper[4735]: E0215 21:08:42.894827 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:08:53 crc kubenswrapper[4735]: I0215 21:08:53.887344 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:08:53 crc kubenswrapper[4735]: E0215 21:08:53.888469 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:09:04 crc kubenswrapper[4735]: I0215 21:09:04.887162 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:09:04 crc kubenswrapper[4735]: E0215 21:09:04.887980 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:09:19 crc kubenswrapper[4735]: I0215 21:09:19.886758 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:09:19 crc kubenswrapper[4735]: E0215 21:09:19.887653 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:09:31 crc kubenswrapper[4735]: I0215 21:09:31.887579 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:09:31 crc kubenswrapper[4735]: E0215 21:09:31.888383 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:09:43 crc kubenswrapper[4735]: I0215 21:09:43.887938 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:09:43 crc kubenswrapper[4735]: E0215 21:09:43.888703 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:09:54 crc kubenswrapper[4735]: I0215 21:09:54.887375 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:09:54 crc kubenswrapper[4735]: E0215 21:09:54.888318 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:10:09 crc kubenswrapper[4735]: I0215 21:10:09.886770 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:10:09 crc kubenswrapper[4735]: E0215 21:10:09.887539 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:10:21 crc kubenswrapper[4735]: I0215 21:10:21.888160 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:10:21 crc kubenswrapper[4735]: E0215 21:10:21.889086 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:10:32 crc kubenswrapper[4735]: I0215 21:10:32.893523 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:10:32 crc kubenswrapper[4735]: E0215 21:10:32.894391 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:10:45 crc kubenswrapper[4735]: I0215 21:10:45.906586 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:10:45 crc kubenswrapper[4735]: E0215 21:10:45.909139 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:10:58 crc kubenswrapper[4735]: I0215 21:10:58.887537 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:10:58 crc kubenswrapper[4735]: E0215 21:10:58.888257 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:11:13 crc kubenswrapper[4735]: I0215 21:11:13.888421 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:11:13 crc kubenswrapper[4735]: E0215 21:11:13.890820 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:11:28 crc kubenswrapper[4735]: I0215 21:11:28.886854 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:11:28 crc kubenswrapper[4735]: E0215 21:11:28.888262 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:11:40 crc kubenswrapper[4735]: I0215 21:11:40.888069 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:11:40 crc kubenswrapper[4735]: E0215 21:11:40.889480 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:11:51 crc kubenswrapper[4735]: I0215 21:11:51.887806 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:11:52 crc kubenswrapper[4735]: I0215 21:11:52.573273 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"7598fe75be9c0b797e578f8109e90db399cb18b220bc8e4eae2fdd497c625fb7"} Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.115228 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-96992"] Feb 15 21:12:13 crc kubenswrapper[4735]: E0215 21:12:13.116151 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerName="registry-server" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.116165 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerName="registry-server" Feb 15 21:12:13 crc kubenswrapper[4735]: E0215 21:12:13.116181 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerName="extract-utilities" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.116189 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerName="extract-utilities" Feb 15 21:12:13 crc kubenswrapper[4735]: E0215 21:12:13.116202 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerName="extract-content" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.116207 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerName="extract-content" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.116381 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fae6e75-eaf2-44c4-b8f1-fa56c642a54c" containerName="registry-server" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.117801 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.132794 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-96992"] Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.188184 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-catalog-content\") pod \"redhat-operators-96992\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.188254 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-utilities\") pod \"redhat-operators-96992\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.188556 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8js7\" (UniqueName: \"kubernetes.io/projected/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-kube-api-access-v8js7\") pod \"redhat-operators-96992\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.289502 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8js7\" (UniqueName: \"kubernetes.io/projected/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-kube-api-access-v8js7\") pod \"redhat-operators-96992\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.289819 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-catalog-content\") pod \"redhat-operators-96992\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.289964 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-utilities\") pod \"redhat-operators-96992\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.290509 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-catalog-content\") pod \"redhat-operators-96992\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.290530 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-utilities\") pod \"redhat-operators-96992\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.310052 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8js7\" (UniqueName: \"kubernetes.io/projected/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-kube-api-access-v8js7\") pod \"redhat-operators-96992\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.438491 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:13 crc kubenswrapper[4735]: I0215 21:12:13.968555 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-96992"] Feb 15 21:12:14 crc kubenswrapper[4735]: I0215 21:12:14.762015 4735 generic.go:334] "Generic (PLEG): container finished" podID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerID="3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394" exitCode=0 Feb 15 21:12:14 crc kubenswrapper[4735]: I0215 21:12:14.762391 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96992" event={"ID":"a2210d5e-730b-4fc4-bd5a-583c9fd398d3","Type":"ContainerDied","Data":"3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394"} Feb 15 21:12:14 crc kubenswrapper[4735]: I0215 21:12:14.762447 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96992" event={"ID":"a2210d5e-730b-4fc4-bd5a-583c9fd398d3","Type":"ContainerStarted","Data":"7c5850fb55e2ed7b4cd19b2c2f002068c0b5dcfe8ba2beda6468abf62afac073"} Feb 15 21:12:14 crc kubenswrapper[4735]: I0215 21:12:14.764597 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 21:12:15 crc kubenswrapper[4735]: I0215 21:12:15.772482 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96992" event={"ID":"a2210d5e-730b-4fc4-bd5a-583c9fd398d3","Type":"ContainerStarted","Data":"92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07"} Feb 15 21:12:20 crc kubenswrapper[4735]: I0215 21:12:20.824364 4735 generic.go:334] "Generic (PLEG): container finished" podID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerID="92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07" exitCode=0 Feb 15 21:12:20 crc kubenswrapper[4735]: I0215 21:12:20.824418 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96992" event={"ID":"a2210d5e-730b-4fc4-bd5a-583c9fd398d3","Type":"ContainerDied","Data":"92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07"} Feb 15 21:12:21 crc kubenswrapper[4735]: I0215 21:12:21.836230 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96992" event={"ID":"a2210d5e-730b-4fc4-bd5a-583c9fd398d3","Type":"ContainerStarted","Data":"69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544"} Feb 15 21:12:21 crc kubenswrapper[4735]: I0215 21:12:21.859500 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-96992" podStartSLOduration=2.384380387 podStartE2EDuration="8.859475584s" podCreationTimestamp="2026-02-15 21:12:13 +0000 UTC" firstStartedPulling="2026-02-15 21:12:14.764403001 +0000 UTC m=+3342.630418624" lastFinishedPulling="2026-02-15 21:12:21.239498198 +0000 UTC m=+3349.105513821" observedRunningTime="2026-02-15 21:12:21.851661774 +0000 UTC m=+3349.717677407" watchObservedRunningTime="2026-02-15 21:12:21.859475584 +0000 UTC m=+3349.725491227" Feb 15 21:12:23 crc kubenswrapper[4735]: I0215 21:12:23.439190 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:23 crc kubenswrapper[4735]: I0215 21:12:23.439260 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:24 crc kubenswrapper[4735]: I0215 21:12:24.491579 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-96992" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="registry-server" probeResult="failure" output=< Feb 15 21:12:24 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:12:24 crc kubenswrapper[4735]: > Feb 15 21:12:34 crc kubenswrapper[4735]: I0215 21:12:34.519769 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-96992" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="registry-server" probeResult="failure" output=< Feb 15 21:12:34 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:12:34 crc kubenswrapper[4735]: > Feb 15 21:12:43 crc kubenswrapper[4735]: I0215 21:12:43.499750 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:43 crc kubenswrapper[4735]: I0215 21:12:43.563985 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:44 crc kubenswrapper[4735]: I0215 21:12:44.337258 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-96992"] Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.025575 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-96992" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="registry-server" containerID="cri-o://69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544" gracePeriod=2 Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.583737 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.652508 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-utilities\") pod \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.652578 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-catalog-content\") pod \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.652599 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8js7\" (UniqueName: \"kubernetes.io/projected/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-kube-api-access-v8js7\") pod \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\" (UID: \"a2210d5e-730b-4fc4-bd5a-583c9fd398d3\") " Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.655330 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-utilities" (OuterVolumeSpecName: "utilities") pod "a2210d5e-730b-4fc4-bd5a-583c9fd398d3" (UID: "a2210d5e-730b-4fc4-bd5a-583c9fd398d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.659478 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-kube-api-access-v8js7" (OuterVolumeSpecName: "kube-api-access-v8js7") pod "a2210d5e-730b-4fc4-bd5a-583c9fd398d3" (UID: "a2210d5e-730b-4fc4-bd5a-583c9fd398d3"). InnerVolumeSpecName "kube-api-access-v8js7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.754683 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.754721 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8js7\" (UniqueName: \"kubernetes.io/projected/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-kube-api-access-v8js7\") on node \"crc\" DevicePath \"\"" Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.766409 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2210d5e-730b-4fc4-bd5a-583c9fd398d3" (UID: "a2210d5e-730b-4fc4-bd5a-583c9fd398d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:12:45 crc kubenswrapper[4735]: I0215 21:12:45.856583 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2210d5e-730b-4fc4-bd5a-583c9fd398d3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.040390 4735 generic.go:334] "Generic (PLEG): container finished" podID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerID="69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544" exitCode=0 Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.040429 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96992" event={"ID":"a2210d5e-730b-4fc4-bd5a-583c9fd398d3","Type":"ContainerDied","Data":"69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544"} Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.040457 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96992" event={"ID":"a2210d5e-730b-4fc4-bd5a-583c9fd398d3","Type":"ContainerDied","Data":"7c5850fb55e2ed7b4cd19b2c2f002068c0b5dcfe8ba2beda6468abf62afac073"} Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.040475 4735 scope.go:117] "RemoveContainer" containerID="69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.040590 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96992" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.074393 4735 scope.go:117] "RemoveContainer" containerID="92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.085325 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-96992"] Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.092542 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-96992"] Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.108790 4735 scope.go:117] "RemoveContainer" containerID="3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.148624 4735 scope.go:117] "RemoveContainer" containerID="69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544" Feb 15 21:12:46 crc kubenswrapper[4735]: E0215 21:12:46.149164 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544\": container with ID starting with 69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544 not found: ID does not exist" containerID="69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.149212 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544"} err="failed to get container status \"69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544\": rpc error: code = NotFound desc = could not find container \"69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544\": container with ID starting with 69ca29ba20532dc173d0f94ef93546f52b4941879cc01538a895ad7d3b1ff544 not found: ID does not exist" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.149239 4735 scope.go:117] "RemoveContainer" containerID="92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07" Feb 15 21:12:46 crc kubenswrapper[4735]: E0215 21:12:46.149579 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07\": container with ID starting with 92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07 not found: ID does not exist" containerID="92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.149630 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07"} err="failed to get container status \"92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07\": rpc error: code = NotFound desc = could not find container \"92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07\": container with ID starting with 92b167abc083629bddb24ba4de44f764cea157d327236cd9bb852b9336764f07 not found: ID does not exist" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.149660 4735 scope.go:117] "RemoveContainer" containerID="3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394" Feb 15 21:12:46 crc kubenswrapper[4735]: E0215 21:12:46.149969 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394\": container with ID starting with 3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394 not found: ID does not exist" containerID="3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.150002 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394"} err="failed to get container status \"3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394\": rpc error: code = NotFound desc = could not find container \"3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394\": container with ID starting with 3dc843649df9b1977e1f35f0ba3b7a674cad589bee29aba4876897aa37ae8394 not found: ID does not exist" Feb 15 21:12:46 crc kubenswrapper[4735]: E0215 21:12:46.160694 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2210d5e_730b_4fc4_bd5a_583c9fd398d3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2210d5e_730b_4fc4_bd5a_583c9fd398d3.slice/crio-7c5850fb55e2ed7b4cd19b2c2f002068c0b5dcfe8ba2beda6468abf62afac073\": RecentStats: unable to find data in memory cache]" Feb 15 21:12:46 crc kubenswrapper[4735]: I0215 21:12:46.897477 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" path="/var/lib/kubelet/pods/a2210d5e-730b-4fc4-bd5a-583c9fd398d3/volumes" Feb 15 21:14:19 crc kubenswrapper[4735]: I0215 21:14:19.679711 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:14:19 crc kubenswrapper[4735]: I0215 21:14:19.680339 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:14:49 crc kubenswrapper[4735]: I0215 21:14:49.680436 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:14:49 crc kubenswrapper[4735]: I0215 21:14:49.681088 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.149550 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w"] Feb 15 21:15:00 crc kubenswrapper[4735]: E0215 21:15:00.150382 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="registry-server" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.150396 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="registry-server" Feb 15 21:15:00 crc kubenswrapper[4735]: E0215 21:15:00.150409 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="extract-content" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.150414 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="extract-content" Feb 15 21:15:00 crc kubenswrapper[4735]: E0215 21:15:00.150440 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="extract-utilities" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.150446 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="extract-utilities" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.150615 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2210d5e-730b-4fc4-bd5a-583c9fd398d3" containerName="registry-server" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.151270 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.155521 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.157226 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.172995 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w"] Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.324822 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv7xv\" (UniqueName: \"kubernetes.io/projected/6524052e-9d12-477e-aafa-9061b2130890-kube-api-access-bv7xv\") pod \"collect-profiles-29519835-9hv6w\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.324877 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6524052e-9d12-477e-aafa-9061b2130890-config-volume\") pod \"collect-profiles-29519835-9hv6w\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.324893 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6524052e-9d12-477e-aafa-9061b2130890-secret-volume\") pod \"collect-profiles-29519835-9hv6w\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.426915 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv7xv\" (UniqueName: \"kubernetes.io/projected/6524052e-9d12-477e-aafa-9061b2130890-kube-api-access-bv7xv\") pod \"collect-profiles-29519835-9hv6w\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.426985 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6524052e-9d12-477e-aafa-9061b2130890-config-volume\") pod \"collect-profiles-29519835-9hv6w\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.427015 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6524052e-9d12-477e-aafa-9061b2130890-secret-volume\") pod \"collect-profiles-29519835-9hv6w\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.429061 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6524052e-9d12-477e-aafa-9061b2130890-config-volume\") pod \"collect-profiles-29519835-9hv6w\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.433237 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6524052e-9d12-477e-aafa-9061b2130890-secret-volume\") pod \"collect-profiles-29519835-9hv6w\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.444923 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv7xv\" (UniqueName: \"kubernetes.io/projected/6524052e-9d12-477e-aafa-9061b2130890-kube-api-access-bv7xv\") pod \"collect-profiles-29519835-9hv6w\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.505131 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:00 crc kubenswrapper[4735]: I0215 21:15:00.950126 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w"] Feb 15 21:15:01 crc kubenswrapper[4735]: I0215 21:15:01.250639 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" event={"ID":"6524052e-9d12-477e-aafa-9061b2130890","Type":"ContainerStarted","Data":"72134e3173a59ed61a1602e2b25f9af421920d0eb0106d03641bf97b4ad546c3"} Feb 15 21:15:01 crc kubenswrapper[4735]: I0215 21:15:01.250968 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" event={"ID":"6524052e-9d12-477e-aafa-9061b2130890","Type":"ContainerStarted","Data":"54749b8aa7a8ef79f9f4c9fb436cfc426ad7a69c72a2c2258b57f66f4678904d"} Feb 15 21:15:01 crc kubenswrapper[4735]: I0215 21:15:01.267926 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" podStartSLOduration=1.267910297 podStartE2EDuration="1.267910297s" podCreationTimestamp="2026-02-15 21:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 21:15:01.264033352 +0000 UTC m=+3509.130048975" watchObservedRunningTime="2026-02-15 21:15:01.267910297 +0000 UTC m=+3509.133925920" Feb 15 21:15:02 crc kubenswrapper[4735]: I0215 21:15:02.264962 4735 generic.go:334] "Generic (PLEG): container finished" podID="6524052e-9d12-477e-aafa-9061b2130890" containerID="72134e3173a59ed61a1602e2b25f9af421920d0eb0106d03641bf97b4ad546c3" exitCode=0 Feb 15 21:15:02 crc kubenswrapper[4735]: I0215 21:15:02.265294 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" event={"ID":"6524052e-9d12-477e-aafa-9061b2130890","Type":"ContainerDied","Data":"72134e3173a59ed61a1602e2b25f9af421920d0eb0106d03641bf97b4ad546c3"} Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.737699 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.895959 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv7xv\" (UniqueName: \"kubernetes.io/projected/6524052e-9d12-477e-aafa-9061b2130890-kube-api-access-bv7xv\") pod \"6524052e-9d12-477e-aafa-9061b2130890\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.896456 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6524052e-9d12-477e-aafa-9061b2130890-config-volume\") pod \"6524052e-9d12-477e-aafa-9061b2130890\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.896753 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6524052e-9d12-477e-aafa-9061b2130890-secret-volume\") pod \"6524052e-9d12-477e-aafa-9061b2130890\" (UID: \"6524052e-9d12-477e-aafa-9061b2130890\") " Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.897083 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6524052e-9d12-477e-aafa-9061b2130890-config-volume" (OuterVolumeSpecName: "config-volume") pod "6524052e-9d12-477e-aafa-9061b2130890" (UID: "6524052e-9d12-477e-aafa-9061b2130890"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.897480 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6524052e-9d12-477e-aafa-9061b2130890-config-volume\") on node \"crc\" DevicePath \"\"" Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.902163 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6524052e-9d12-477e-aafa-9061b2130890-kube-api-access-bv7xv" (OuterVolumeSpecName: "kube-api-access-bv7xv") pod "6524052e-9d12-477e-aafa-9061b2130890" (UID: "6524052e-9d12-477e-aafa-9061b2130890"). InnerVolumeSpecName "kube-api-access-bv7xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.903189 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6524052e-9d12-477e-aafa-9061b2130890-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6524052e-9d12-477e-aafa-9061b2130890" (UID: "6524052e-9d12-477e-aafa-9061b2130890"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.999570 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6524052e-9d12-477e-aafa-9061b2130890-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 15 21:15:03 crc kubenswrapper[4735]: I0215 21:15:03.999837 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv7xv\" (UniqueName: \"kubernetes.io/projected/6524052e-9d12-477e-aafa-9061b2130890-kube-api-access-bv7xv\") on node \"crc\" DevicePath \"\"" Feb 15 21:15:04 crc kubenswrapper[4735]: I0215 21:15:04.288400 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" event={"ID":"6524052e-9d12-477e-aafa-9061b2130890","Type":"ContainerDied","Data":"54749b8aa7a8ef79f9f4c9fb436cfc426ad7a69c72a2c2258b57f66f4678904d"} Feb 15 21:15:04 crc kubenswrapper[4735]: I0215 21:15:04.288438 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54749b8aa7a8ef79f9f4c9fb436cfc426ad7a69c72a2c2258b57f66f4678904d" Feb 15 21:15:04 crc kubenswrapper[4735]: I0215 21:15:04.288486 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519835-9hv6w" Feb 15 21:15:04 crc kubenswrapper[4735]: I0215 21:15:04.369263 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr"] Feb 15 21:15:04 crc kubenswrapper[4735]: I0215 21:15:04.376690 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519790-dkjpr"] Feb 15 21:15:04 crc kubenswrapper[4735]: I0215 21:15:04.911621 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab473fa2-1b7e-4ea8-ac98-8d883266e2ca" path="/var/lib/kubelet/pods/ab473fa2-1b7e-4ea8-ac98-8d883266e2ca/volumes" Feb 15 21:15:16 crc kubenswrapper[4735]: I0215 21:15:16.910323 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jrzhf"] Feb 15 21:15:16 crc kubenswrapper[4735]: E0215 21:15:16.911378 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6524052e-9d12-477e-aafa-9061b2130890" containerName="collect-profiles" Feb 15 21:15:16 crc kubenswrapper[4735]: I0215 21:15:16.911394 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="6524052e-9d12-477e-aafa-9061b2130890" containerName="collect-profiles" Feb 15 21:15:16 crc kubenswrapper[4735]: I0215 21:15:16.911646 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="6524052e-9d12-477e-aafa-9061b2130890" containerName="collect-profiles" Feb 15 21:15:16 crc kubenswrapper[4735]: I0215 21:15:16.913559 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:16 crc kubenswrapper[4735]: I0215 21:15:16.962080 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jrzhf"] Feb 15 21:15:16 crc kubenswrapper[4735]: I0215 21:15:16.963906 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-catalog-content\") pod \"redhat-marketplace-jrzhf\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:16 crc kubenswrapper[4735]: I0215 21:15:16.963930 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-utilities\") pod \"redhat-marketplace-jrzhf\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:16 crc kubenswrapper[4735]: I0215 21:15:16.964054 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mqqd\" (UniqueName: \"kubernetes.io/projected/dc4cf06c-2639-44b0-b121-87d762d4a999-kube-api-access-8mqqd\") pod \"redhat-marketplace-jrzhf\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:17 crc kubenswrapper[4735]: I0215 21:15:17.066241 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-catalog-content\") pod \"redhat-marketplace-jrzhf\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:17 crc kubenswrapper[4735]: I0215 21:15:17.066276 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-utilities\") pod \"redhat-marketplace-jrzhf\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:17 crc kubenswrapper[4735]: I0215 21:15:17.066354 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mqqd\" (UniqueName: \"kubernetes.io/projected/dc4cf06c-2639-44b0-b121-87d762d4a999-kube-api-access-8mqqd\") pod \"redhat-marketplace-jrzhf\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:17 crc kubenswrapper[4735]: I0215 21:15:17.066782 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-catalog-content\") pod \"redhat-marketplace-jrzhf\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:17 crc kubenswrapper[4735]: I0215 21:15:17.066974 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-utilities\") pod \"redhat-marketplace-jrzhf\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:17 crc kubenswrapper[4735]: I0215 21:15:17.094994 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mqqd\" (UniqueName: \"kubernetes.io/projected/dc4cf06c-2639-44b0-b121-87d762d4a999-kube-api-access-8mqqd\") pod \"redhat-marketplace-jrzhf\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:17 crc kubenswrapper[4735]: I0215 21:15:17.284773 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:17 crc kubenswrapper[4735]: I0215 21:15:17.830359 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jrzhf"] Feb 15 21:15:18 crc kubenswrapper[4735]: I0215 21:15:18.488321 4735 generic.go:334] "Generic (PLEG): container finished" podID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerID="9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640" exitCode=0 Feb 15 21:15:18 crc kubenswrapper[4735]: I0215 21:15:18.488389 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jrzhf" event={"ID":"dc4cf06c-2639-44b0-b121-87d762d4a999","Type":"ContainerDied","Data":"9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640"} Feb 15 21:15:18 crc kubenswrapper[4735]: I0215 21:15:18.488637 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jrzhf" event={"ID":"dc4cf06c-2639-44b0-b121-87d762d4a999","Type":"ContainerStarted","Data":"a5ca03ee50fb59f48e8ba26cc7241a030df71d9ab94677c8ff5ec04b161dee65"} Feb 15 21:15:19 crc kubenswrapper[4735]: I0215 21:15:19.501200 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jrzhf" event={"ID":"dc4cf06c-2639-44b0-b121-87d762d4a999","Type":"ContainerStarted","Data":"9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f"} Feb 15 21:15:19 crc kubenswrapper[4735]: I0215 21:15:19.679447 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:15:19 crc kubenswrapper[4735]: I0215 21:15:19.679507 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:15:19 crc kubenswrapper[4735]: I0215 21:15:19.679554 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 21:15:19 crc kubenswrapper[4735]: I0215 21:15:19.680267 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7598fe75be9c0b797e578f8109e90db399cb18b220bc8e4eae2fdd497c625fb7"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 21:15:19 crc kubenswrapper[4735]: I0215 21:15:19.680327 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://7598fe75be9c0b797e578f8109e90db399cb18b220bc8e4eae2fdd497c625fb7" gracePeriod=600 Feb 15 21:15:19 crc kubenswrapper[4735]: E0215 21:15:19.789850 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43de4639_a922_4182_9377_d1f28b3aa4c1.slice/crio-7598fe75be9c0b797e578f8109e90db399cb18b220bc8e4eae2fdd497c625fb7.scope\": RecentStats: unable to find data in memory cache]" Feb 15 21:15:20 crc kubenswrapper[4735]: I0215 21:15:20.512347 4735 generic.go:334] "Generic (PLEG): container finished" podID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerID="9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f" exitCode=0 Feb 15 21:15:20 crc kubenswrapper[4735]: I0215 21:15:20.512445 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jrzhf" event={"ID":"dc4cf06c-2639-44b0-b121-87d762d4a999","Type":"ContainerDied","Data":"9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f"} Feb 15 21:15:20 crc kubenswrapper[4735]: I0215 21:15:20.520023 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="7598fe75be9c0b797e578f8109e90db399cb18b220bc8e4eae2fdd497c625fb7" exitCode=0 Feb 15 21:15:20 crc kubenswrapper[4735]: I0215 21:15:20.520085 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"7598fe75be9c0b797e578f8109e90db399cb18b220bc8e4eae2fdd497c625fb7"} Feb 15 21:15:20 crc kubenswrapper[4735]: I0215 21:15:20.520111 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca"} Feb 15 21:15:20 crc kubenswrapper[4735]: I0215 21:15:20.520127 4735 scope.go:117] "RemoveContainer" containerID="422462d673db69c8b9ba020e35ea37ca18e78f5a8a0207ceeee23e4816e931e4" Feb 15 21:15:21 crc kubenswrapper[4735]: I0215 21:15:21.534087 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jrzhf" event={"ID":"dc4cf06c-2639-44b0-b121-87d762d4a999","Type":"ContainerStarted","Data":"2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917"} Feb 15 21:15:21 crc kubenswrapper[4735]: I0215 21:15:21.553422 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jrzhf" podStartSLOduration=3.115409148 podStartE2EDuration="5.553406477s" podCreationTimestamp="2026-02-15 21:15:16 +0000 UTC" firstStartedPulling="2026-02-15 21:15:18.490363896 +0000 UTC m=+3526.356379519" lastFinishedPulling="2026-02-15 21:15:20.928361225 +0000 UTC m=+3528.794376848" observedRunningTime="2026-02-15 21:15:21.551228048 +0000 UTC m=+3529.417243661" watchObservedRunningTime="2026-02-15 21:15:21.553406477 +0000 UTC m=+3529.419422100" Feb 15 21:15:27 crc kubenswrapper[4735]: I0215 21:15:27.286150 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:27 crc kubenswrapper[4735]: I0215 21:15:27.286619 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:27 crc kubenswrapper[4735]: I0215 21:15:27.334829 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:27 crc kubenswrapper[4735]: I0215 21:15:27.743561 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:27 crc kubenswrapper[4735]: I0215 21:15:27.795895 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jrzhf"] Feb 15 21:15:29 crc kubenswrapper[4735]: I0215 21:15:29.616267 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jrzhf" podUID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerName="registry-server" containerID="cri-o://2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917" gracePeriod=2 Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.113684 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.282641 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-utilities\") pod \"dc4cf06c-2639-44b0-b121-87d762d4a999\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.282756 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mqqd\" (UniqueName: \"kubernetes.io/projected/dc4cf06c-2639-44b0-b121-87d762d4a999-kube-api-access-8mqqd\") pod \"dc4cf06c-2639-44b0-b121-87d762d4a999\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.282801 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-catalog-content\") pod \"dc4cf06c-2639-44b0-b121-87d762d4a999\" (UID: \"dc4cf06c-2639-44b0-b121-87d762d4a999\") " Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.283423 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-utilities" (OuterVolumeSpecName: "utilities") pod "dc4cf06c-2639-44b0-b121-87d762d4a999" (UID: "dc4cf06c-2639-44b0-b121-87d762d4a999"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.293878 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc4cf06c-2639-44b0-b121-87d762d4a999-kube-api-access-8mqqd" (OuterVolumeSpecName: "kube-api-access-8mqqd") pod "dc4cf06c-2639-44b0-b121-87d762d4a999" (UID: "dc4cf06c-2639-44b0-b121-87d762d4a999"). InnerVolumeSpecName "kube-api-access-8mqqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.305331 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc4cf06c-2639-44b0-b121-87d762d4a999" (UID: "dc4cf06c-2639-44b0-b121-87d762d4a999"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.385180 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.385231 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mqqd\" (UniqueName: \"kubernetes.io/projected/dc4cf06c-2639-44b0-b121-87d762d4a999-kube-api-access-8mqqd\") on node \"crc\" DevicePath \"\"" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.385246 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4cf06c-2639-44b0-b121-87d762d4a999-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.632096 4735 generic.go:334] "Generic (PLEG): container finished" podID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerID="2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917" exitCode=0 Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.632137 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jrzhf" event={"ID":"dc4cf06c-2639-44b0-b121-87d762d4a999","Type":"ContainerDied","Data":"2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917"} Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.632161 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jrzhf" event={"ID":"dc4cf06c-2639-44b0-b121-87d762d4a999","Type":"ContainerDied","Data":"a5ca03ee50fb59f48e8ba26cc7241a030df71d9ab94677c8ff5ec04b161dee65"} Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.632177 4735 scope.go:117] "RemoveContainer" containerID="2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.632309 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jrzhf" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.661912 4735 scope.go:117] "RemoveContainer" containerID="9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.683003 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jrzhf"] Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.694096 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jrzhf"] Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.707995 4735 scope.go:117] "RemoveContainer" containerID="9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.754600 4735 scope.go:117] "RemoveContainer" containerID="2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917" Feb 15 21:15:30 crc kubenswrapper[4735]: E0215 21:15:30.755121 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917\": container with ID starting with 2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917 not found: ID does not exist" containerID="2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.755150 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917"} err="failed to get container status \"2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917\": rpc error: code = NotFound desc = could not find container \"2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917\": container with ID starting with 2c38155a9784057a9ae369024a5abd8a41d458b9761940f3cfa608a5a9e51917 not found: ID does not exist" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.755169 4735 scope.go:117] "RemoveContainer" containerID="9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f" Feb 15 21:15:30 crc kubenswrapper[4735]: E0215 21:15:30.755561 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f\": container with ID starting with 9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f not found: ID does not exist" containerID="9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.755580 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f"} err="failed to get container status \"9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f\": rpc error: code = NotFound desc = could not find container \"9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f\": container with ID starting with 9fbc6c5712c411482fa710127f12829439c589e7973d9996108ff059fbd3a98f not found: ID does not exist" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.755619 4735 scope.go:117] "RemoveContainer" containerID="9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640" Feb 15 21:15:30 crc kubenswrapper[4735]: E0215 21:15:30.755830 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640\": container with ID starting with 9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640 not found: ID does not exist" containerID="9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.755851 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640"} err="failed to get container status \"9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640\": rpc error: code = NotFound desc = could not find container \"9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640\": container with ID starting with 9ae2d4626b3e4226194eaa9d129b8da3e29c57863393febb331c15deacaad640 not found: ID does not exist" Feb 15 21:15:30 crc kubenswrapper[4735]: I0215 21:15:30.898703 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc4cf06c-2639-44b0-b121-87d762d4a999" path="/var/lib/kubelet/pods/dc4cf06c-2639-44b0-b121-87d762d4a999/volumes" Feb 15 21:15:53 crc kubenswrapper[4735]: I0215 21:15:53.033244 4735 scope.go:117] "RemoveContainer" containerID="fa0e6b2b0b04222955ece57389214bb43f2762b8ef5b9fd83c15651879bf9ede" Feb 15 21:17:49 crc kubenswrapper[4735]: I0215 21:17:49.680418 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:17:49 crc kubenswrapper[4735]: I0215 21:17:49.680976 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:18:19 crc kubenswrapper[4735]: I0215 21:18:19.680276 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:18:19 crc kubenswrapper[4735]: I0215 21:18:19.680850 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:18:49 crc kubenswrapper[4735]: I0215 21:18:49.680346 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:18:49 crc kubenswrapper[4735]: I0215 21:18:49.680815 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:18:49 crc kubenswrapper[4735]: I0215 21:18:49.680855 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 21:18:49 crc kubenswrapper[4735]: I0215 21:18:49.681603 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 21:18:49 crc kubenswrapper[4735]: I0215 21:18:49.681651 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" gracePeriod=600 Feb 15 21:18:49 crc kubenswrapper[4735]: E0215 21:18:49.802495 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:18:50 crc kubenswrapper[4735]: I0215 21:18:50.439763 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" exitCode=0 Feb 15 21:18:50 crc kubenswrapper[4735]: I0215 21:18:50.439793 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca"} Feb 15 21:18:50 crc kubenswrapper[4735]: I0215 21:18:50.440155 4735 scope.go:117] "RemoveContainer" containerID="7598fe75be9c0b797e578f8109e90db399cb18b220bc8e4eae2fdd497c625fb7" Feb 15 21:18:50 crc kubenswrapper[4735]: I0215 21:18:50.441369 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:18:50 crc kubenswrapper[4735]: E0215 21:18:50.441876 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:19:01 crc kubenswrapper[4735]: I0215 21:19:01.886886 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:19:01 crc kubenswrapper[4735]: E0215 21:19:01.887777 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:19:15 crc kubenswrapper[4735]: I0215 21:19:15.886531 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:19:15 crc kubenswrapper[4735]: E0215 21:19:15.887243 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:19:28 crc kubenswrapper[4735]: I0215 21:19:28.887229 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:19:28 crc kubenswrapper[4735]: E0215 21:19:28.888222 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:19:40 crc kubenswrapper[4735]: I0215 21:19:40.886868 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:19:40 crc kubenswrapper[4735]: E0215 21:19:40.887718 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:19:52 crc kubenswrapper[4735]: I0215 21:19:52.895665 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:19:52 crc kubenswrapper[4735]: E0215 21:19:52.896417 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:20:05 crc kubenswrapper[4735]: I0215 21:20:05.886366 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:20:05 crc kubenswrapper[4735]: E0215 21:20:05.887161 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:20:17 crc kubenswrapper[4735]: I0215 21:20:17.888661 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:20:17 crc kubenswrapper[4735]: E0215 21:20:17.889734 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:20:31 crc kubenswrapper[4735]: I0215 21:20:31.886527 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:20:31 crc kubenswrapper[4735]: E0215 21:20:31.887336 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:20:42 crc kubenswrapper[4735]: I0215 21:20:42.893178 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:20:42 crc kubenswrapper[4735]: E0215 21:20:42.894798 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:20:57 crc kubenswrapper[4735]: I0215 21:20:57.887752 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:20:57 crc kubenswrapper[4735]: E0215 21:20:57.888941 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:21:11 crc kubenswrapper[4735]: I0215 21:21:11.887181 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:21:11 crc kubenswrapper[4735]: E0215 21:21:11.888033 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:21:24 crc kubenswrapper[4735]: I0215 21:21:24.887352 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:21:24 crc kubenswrapper[4735]: E0215 21:21:24.888363 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:21:35 crc kubenswrapper[4735]: I0215 21:21:35.887216 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:21:35 crc kubenswrapper[4735]: E0215 21:21:35.888791 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:21:46 crc kubenswrapper[4735]: I0215 21:21:46.887978 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:21:46 crc kubenswrapper[4735]: E0215 21:21:46.889339 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.354722 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zvn67"] Feb 15 21:21:54 crc kubenswrapper[4735]: E0215 21:21:54.355574 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerName="extract-content" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.355588 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerName="extract-content" Feb 15 21:21:54 crc kubenswrapper[4735]: E0215 21:21:54.355601 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerName="extract-utilities" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.355620 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerName="extract-utilities" Feb 15 21:21:54 crc kubenswrapper[4735]: E0215 21:21:54.355659 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerName="registry-server" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.355665 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerName="registry-server" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.355884 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4cf06c-2639-44b0-b121-87d762d4a999" containerName="registry-server" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.357242 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.371126 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvn67"] Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.444656 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-catalog-content\") pod \"certified-operators-zvn67\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.444736 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8b47\" (UniqueName: \"kubernetes.io/projected/f380e363-4c49-4dfd-ab68-214b4c87015f-kube-api-access-q8b47\") pod \"certified-operators-zvn67\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.444844 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-utilities\") pod \"certified-operators-zvn67\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.546830 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8b47\" (UniqueName: \"kubernetes.io/projected/f380e363-4c49-4dfd-ab68-214b4c87015f-kube-api-access-q8b47\") pod \"certified-operators-zvn67\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.546915 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-utilities\") pod \"certified-operators-zvn67\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.547033 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-catalog-content\") pod \"certified-operators-zvn67\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.547446 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-catalog-content\") pod \"certified-operators-zvn67\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.547716 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-utilities\") pod \"certified-operators-zvn67\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.552500 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nsndx"] Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.560358 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.587838 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nsndx"] Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.596641 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8b47\" (UniqueName: \"kubernetes.io/projected/f380e363-4c49-4dfd-ab68-214b4c87015f-kube-api-access-q8b47\") pod \"certified-operators-zvn67\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.648565 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-utilities\") pod \"community-operators-nsndx\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.648647 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-catalog-content\") pod \"community-operators-nsndx\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.648809 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-875b6\" (UniqueName: \"kubernetes.io/projected/70d6d340-66ed-43cc-9971-bee82135081d-kube-api-access-875b6\") pod \"community-operators-nsndx\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.680700 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.751189 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-875b6\" (UniqueName: \"kubernetes.io/projected/70d6d340-66ed-43cc-9971-bee82135081d-kube-api-access-875b6\") pod \"community-operators-nsndx\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.751539 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-utilities\") pod \"community-operators-nsndx\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.751583 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-catalog-content\") pod \"community-operators-nsndx\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.751985 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-catalog-content\") pod \"community-operators-nsndx\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.752099 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-utilities\") pod \"community-operators-nsndx\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.779937 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-875b6\" (UniqueName: \"kubernetes.io/projected/70d6d340-66ed-43cc-9971-bee82135081d-kube-api-access-875b6\") pod \"community-operators-nsndx\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:54 crc kubenswrapper[4735]: I0215 21:21:54.948171 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:21:55 crc kubenswrapper[4735]: I0215 21:21:55.194379 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvn67"] Feb 15 21:21:55 crc kubenswrapper[4735]: W0215 21:21:55.210759 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf380e363_4c49_4dfd_ab68_214b4c87015f.slice/crio-bf0593cddedb0bf470db3cfd4e8895aef645695c2017393ebe8675361a98b549 WatchSource:0}: Error finding container bf0593cddedb0bf470db3cfd4e8895aef645695c2017393ebe8675361a98b549: Status 404 returned error can't find the container with id bf0593cddedb0bf470db3cfd4e8895aef645695c2017393ebe8675361a98b549 Feb 15 21:21:55 crc kubenswrapper[4735]: I0215 21:21:55.413362 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvn67" event={"ID":"f380e363-4c49-4dfd-ab68-214b4c87015f","Type":"ContainerStarted","Data":"bf0593cddedb0bf470db3cfd4e8895aef645695c2017393ebe8675361a98b549"} Feb 15 21:21:55 crc kubenswrapper[4735]: I0215 21:21:55.707175 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nsndx"] Feb 15 21:21:56 crc kubenswrapper[4735]: I0215 21:21:56.422974 4735 generic.go:334] "Generic (PLEG): container finished" podID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerID="e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189" exitCode=0 Feb 15 21:21:56 crc kubenswrapper[4735]: I0215 21:21:56.423079 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvn67" event={"ID":"f380e363-4c49-4dfd-ab68-214b4c87015f","Type":"ContainerDied","Data":"e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189"} Feb 15 21:21:56 crc kubenswrapper[4735]: I0215 21:21:56.425070 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 21:21:56 crc kubenswrapper[4735]: I0215 21:21:56.425692 4735 generic.go:334] "Generic (PLEG): container finished" podID="70d6d340-66ed-43cc-9971-bee82135081d" containerID="1872e5fde2e7ccbd1d71bfa04d24c8faefdb7b308f4b2063e07fed65999484ba" exitCode=0 Feb 15 21:21:56 crc kubenswrapper[4735]: I0215 21:21:56.425746 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsndx" event={"ID":"70d6d340-66ed-43cc-9971-bee82135081d","Type":"ContainerDied","Data":"1872e5fde2e7ccbd1d71bfa04d24c8faefdb7b308f4b2063e07fed65999484ba"} Feb 15 21:21:56 crc kubenswrapper[4735]: I0215 21:21:56.425781 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsndx" event={"ID":"70d6d340-66ed-43cc-9971-bee82135081d","Type":"ContainerStarted","Data":"66219a233504493c02975a2663286b6ace614428a152b7bb2256daead6eeec90"} Feb 15 21:21:57 crc kubenswrapper[4735]: I0215 21:21:57.435244 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvn67" event={"ID":"f380e363-4c49-4dfd-ab68-214b4c87015f","Type":"ContainerStarted","Data":"15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f"} Feb 15 21:21:57 crc kubenswrapper[4735]: I0215 21:21:57.438104 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsndx" event={"ID":"70d6d340-66ed-43cc-9971-bee82135081d","Type":"ContainerStarted","Data":"f444aba978af7a8f6d68c6e6dee8235c13d511e528b3cb88666f83a2829522cd"} Feb 15 21:21:59 crc kubenswrapper[4735]: I0215 21:21:59.461829 4735 generic.go:334] "Generic (PLEG): container finished" podID="70d6d340-66ed-43cc-9971-bee82135081d" containerID="f444aba978af7a8f6d68c6e6dee8235c13d511e528b3cb88666f83a2829522cd" exitCode=0 Feb 15 21:21:59 crc kubenswrapper[4735]: I0215 21:21:59.462011 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsndx" event={"ID":"70d6d340-66ed-43cc-9971-bee82135081d","Type":"ContainerDied","Data":"f444aba978af7a8f6d68c6e6dee8235c13d511e528b3cb88666f83a2829522cd"} Feb 15 21:21:59 crc kubenswrapper[4735]: I0215 21:21:59.478568 4735 generic.go:334] "Generic (PLEG): container finished" podID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerID="15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f" exitCode=0 Feb 15 21:21:59 crc kubenswrapper[4735]: I0215 21:21:59.478653 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvn67" event={"ID":"f380e363-4c49-4dfd-ab68-214b4c87015f","Type":"ContainerDied","Data":"15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f"} Feb 15 21:22:00 crc kubenswrapper[4735]: I0215 21:22:00.486774 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvn67" event={"ID":"f380e363-4c49-4dfd-ab68-214b4c87015f","Type":"ContainerStarted","Data":"438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44"} Feb 15 21:22:00 crc kubenswrapper[4735]: I0215 21:22:00.489099 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsndx" event={"ID":"70d6d340-66ed-43cc-9971-bee82135081d","Type":"ContainerStarted","Data":"24c5edc41388241628f9f61b1e3d67386b0b2c29b4f26395e4f8853cd6de26a0"} Feb 15 21:22:00 crc kubenswrapper[4735]: I0215 21:22:00.510291 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zvn67" podStartSLOduration=2.949877735 podStartE2EDuration="6.510262395s" podCreationTimestamp="2026-02-15 21:21:54 +0000 UTC" firstStartedPulling="2026-02-15 21:21:56.424844596 +0000 UTC m=+3924.290860209" lastFinishedPulling="2026-02-15 21:21:59.985229246 +0000 UTC m=+3927.851244869" observedRunningTime="2026-02-15 21:22:00.505387934 +0000 UTC m=+3928.371403557" watchObservedRunningTime="2026-02-15 21:22:00.510262395 +0000 UTC m=+3928.376278018" Feb 15 21:22:00 crc kubenswrapper[4735]: I0215 21:22:00.526816 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nsndx" podStartSLOduration=3.10757902 podStartE2EDuration="6.526798139s" podCreationTimestamp="2026-02-15 21:21:54 +0000 UTC" firstStartedPulling="2026-02-15 21:21:56.426711926 +0000 UTC m=+3924.292727549" lastFinishedPulling="2026-02-15 21:21:59.845931045 +0000 UTC m=+3927.711946668" observedRunningTime="2026-02-15 21:22:00.52646332 +0000 UTC m=+3928.392478943" watchObservedRunningTime="2026-02-15 21:22:00.526798139 +0000 UTC m=+3928.392813762" Feb 15 21:22:00 crc kubenswrapper[4735]: I0215 21:22:00.886770 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:22:00 crc kubenswrapper[4735]: E0215 21:22:00.887017 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:22:04 crc kubenswrapper[4735]: I0215 21:22:04.681786 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:22:04 crc kubenswrapper[4735]: I0215 21:22:04.682274 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:22:04 crc kubenswrapper[4735]: I0215 21:22:04.949039 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:22:04 crc kubenswrapper[4735]: I0215 21:22:04.949414 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:22:05 crc kubenswrapper[4735]: I0215 21:22:05.727009 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-zvn67" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerName="registry-server" probeResult="failure" output=< Feb 15 21:22:05 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:22:05 crc kubenswrapper[4735]: > Feb 15 21:22:05 crc kubenswrapper[4735]: I0215 21:22:05.990647 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-nsndx" podUID="70d6d340-66ed-43cc-9971-bee82135081d" containerName="registry-server" probeResult="failure" output=< Feb 15 21:22:05 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:22:05 crc kubenswrapper[4735]: > Feb 15 21:22:14 crc kubenswrapper[4735]: I0215 21:22:14.725914 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:22:14 crc kubenswrapper[4735]: I0215 21:22:14.809807 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:22:14 crc kubenswrapper[4735]: I0215 21:22:14.935233 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:22:14 crc kubenswrapper[4735]: E0215 21:22:14.935496 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:22:14 crc kubenswrapper[4735]: I0215 21:22:14.985068 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zvn67"] Feb 15 21:22:15 crc kubenswrapper[4735]: I0215 21:22:15.013743 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:22:15 crc kubenswrapper[4735]: I0215 21:22:15.074408 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:22:16 crc kubenswrapper[4735]: I0215 21:22:16.623584 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zvn67" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerName="registry-server" containerID="cri-o://438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44" gracePeriod=2 Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.121540 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.239237 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-catalog-content\") pod \"f380e363-4c49-4dfd-ab68-214b4c87015f\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.239344 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8b47\" (UniqueName: \"kubernetes.io/projected/f380e363-4c49-4dfd-ab68-214b4c87015f-kube-api-access-q8b47\") pod \"f380e363-4c49-4dfd-ab68-214b4c87015f\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.239513 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-utilities\") pod \"f380e363-4c49-4dfd-ab68-214b4c87015f\" (UID: \"f380e363-4c49-4dfd-ab68-214b4c87015f\") " Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.240426 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-utilities" (OuterVolumeSpecName: "utilities") pod "f380e363-4c49-4dfd-ab68-214b4c87015f" (UID: "f380e363-4c49-4dfd-ab68-214b4c87015f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.248805 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f380e363-4c49-4dfd-ab68-214b4c87015f-kube-api-access-q8b47" (OuterVolumeSpecName: "kube-api-access-q8b47") pod "f380e363-4c49-4dfd-ab68-214b4c87015f" (UID: "f380e363-4c49-4dfd-ab68-214b4c87015f"). InnerVolumeSpecName "kube-api-access-q8b47". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.285826 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f380e363-4c49-4dfd-ab68-214b4c87015f" (UID: "f380e363-4c49-4dfd-ab68-214b4c87015f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.341072 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.341287 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f380e363-4c49-4dfd-ab68-214b4c87015f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.341354 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8b47\" (UniqueName: \"kubernetes.io/projected/f380e363-4c49-4dfd-ab68-214b4c87015f-kube-api-access-q8b47\") on node \"crc\" DevicePath \"\"" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.366908 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nsndx"] Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.367202 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nsndx" podUID="70d6d340-66ed-43cc-9971-bee82135081d" containerName="registry-server" containerID="cri-o://24c5edc41388241628f9f61b1e3d67386b0b2c29b4f26395e4f8853cd6de26a0" gracePeriod=2 Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.637063 4735 generic.go:334] "Generic (PLEG): container finished" podID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerID="438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44" exitCode=0 Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.637132 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvn67" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.637143 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvn67" event={"ID":"f380e363-4c49-4dfd-ab68-214b4c87015f","Type":"ContainerDied","Data":"438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44"} Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.637173 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvn67" event={"ID":"f380e363-4c49-4dfd-ab68-214b4c87015f","Type":"ContainerDied","Data":"bf0593cddedb0bf470db3cfd4e8895aef645695c2017393ebe8675361a98b549"} Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.637193 4735 scope.go:117] "RemoveContainer" containerID="438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.642956 4735 generic.go:334] "Generic (PLEG): container finished" podID="70d6d340-66ed-43cc-9971-bee82135081d" containerID="24c5edc41388241628f9f61b1e3d67386b0b2c29b4f26395e4f8853cd6de26a0" exitCode=0 Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.643058 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsndx" event={"ID":"70d6d340-66ed-43cc-9971-bee82135081d","Type":"ContainerDied","Data":"24c5edc41388241628f9f61b1e3d67386b0b2c29b4f26395e4f8853cd6de26a0"} Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.681642 4735 scope.go:117] "RemoveContainer" containerID="15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.695479 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zvn67"] Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.713895 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zvn67"] Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.728627 4735 scope.go:117] "RemoveContainer" containerID="e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.799957 4735 scope.go:117] "RemoveContainer" containerID="438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44" Feb 15 21:22:17 crc kubenswrapper[4735]: E0215 21:22:17.800358 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44\": container with ID starting with 438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44 not found: ID does not exist" containerID="438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.800422 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44"} err="failed to get container status \"438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44\": rpc error: code = NotFound desc = could not find container \"438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44\": container with ID starting with 438123aea527950ff8c372ab7a2146a791043c09c703cfc8e35fd5a1d8f77b44 not found: ID does not exist" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.800443 4735 scope.go:117] "RemoveContainer" containerID="15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f" Feb 15 21:22:17 crc kubenswrapper[4735]: E0215 21:22:17.800771 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f\": container with ID starting with 15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f not found: ID does not exist" containerID="15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.800816 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f"} err="failed to get container status \"15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f\": rpc error: code = NotFound desc = could not find container \"15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f\": container with ID starting with 15a8edef703d4874555a32169ac93732b6d10dc04352687e511feb78f837500f not found: ID does not exist" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.800850 4735 scope.go:117] "RemoveContainer" containerID="e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189" Feb 15 21:22:17 crc kubenswrapper[4735]: E0215 21:22:17.801114 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189\": container with ID starting with e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189 not found: ID does not exist" containerID="e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.801140 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189"} err="failed to get container status \"e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189\": rpc error: code = NotFound desc = could not find container \"e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189\": container with ID starting with e90c88738565f89d0fb5205d9d8dced2a7f65427195e2d45b4ab5236924a9189 not found: ID does not exist" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.884063 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.962292 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-875b6\" (UniqueName: \"kubernetes.io/projected/70d6d340-66ed-43cc-9971-bee82135081d-kube-api-access-875b6\") pod \"70d6d340-66ed-43cc-9971-bee82135081d\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.962362 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-utilities\") pod \"70d6d340-66ed-43cc-9971-bee82135081d\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.962441 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-catalog-content\") pod \"70d6d340-66ed-43cc-9971-bee82135081d\" (UID: \"70d6d340-66ed-43cc-9971-bee82135081d\") " Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.963285 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-utilities" (OuterVolumeSpecName: "utilities") pod "70d6d340-66ed-43cc-9971-bee82135081d" (UID: "70d6d340-66ed-43cc-9971-bee82135081d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:22:17 crc kubenswrapper[4735]: I0215 21:22:17.967233 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70d6d340-66ed-43cc-9971-bee82135081d-kube-api-access-875b6" (OuterVolumeSpecName: "kube-api-access-875b6") pod "70d6d340-66ed-43cc-9971-bee82135081d" (UID: "70d6d340-66ed-43cc-9971-bee82135081d"). InnerVolumeSpecName "kube-api-access-875b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.011926 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70d6d340-66ed-43cc-9971-bee82135081d" (UID: "70d6d340-66ed-43cc-9971-bee82135081d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.064823 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-875b6\" (UniqueName: \"kubernetes.io/projected/70d6d340-66ed-43cc-9971-bee82135081d-kube-api-access-875b6\") on node \"crc\" DevicePath \"\"" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.064860 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.064874 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d6d340-66ed-43cc-9971-bee82135081d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.653579 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsndx" event={"ID":"70d6d340-66ed-43cc-9971-bee82135081d","Type":"ContainerDied","Data":"66219a233504493c02975a2663286b6ace614428a152b7bb2256daead6eeec90"} Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.653630 4735 scope.go:117] "RemoveContainer" containerID="24c5edc41388241628f9f61b1e3d67386b0b2c29b4f26395e4f8853cd6de26a0" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.653668 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsndx" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.676382 4735 scope.go:117] "RemoveContainer" containerID="f444aba978af7a8f6d68c6e6dee8235c13d511e528b3cb88666f83a2829522cd" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.690010 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nsndx"] Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.697177 4735 scope.go:117] "RemoveContainer" containerID="1872e5fde2e7ccbd1d71bfa04d24c8faefdb7b308f4b2063e07fed65999484ba" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.699316 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nsndx"] Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.898121 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70d6d340-66ed-43cc-9971-bee82135081d" path="/var/lib/kubelet/pods/70d6d340-66ed-43cc-9971-bee82135081d/volumes" Feb 15 21:22:18 crc kubenswrapper[4735]: I0215 21:22:18.898921 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" path="/var/lib/kubelet/pods/f380e363-4c49-4dfd-ab68-214b4c87015f/volumes" Feb 15 21:22:28 crc kubenswrapper[4735]: I0215 21:22:28.887436 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:22:28 crc kubenswrapper[4735]: E0215 21:22:28.888110 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.677753 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l92ph"] Feb 15 21:22:40 crc kubenswrapper[4735]: E0215 21:22:40.678640 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d6d340-66ed-43cc-9971-bee82135081d" containerName="extract-content" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.678652 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d6d340-66ed-43cc-9971-bee82135081d" containerName="extract-content" Feb 15 21:22:40 crc kubenswrapper[4735]: E0215 21:22:40.678672 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerName="registry-server" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.678678 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerName="registry-server" Feb 15 21:22:40 crc kubenswrapper[4735]: E0215 21:22:40.678701 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerName="extract-utilities" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.678707 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerName="extract-utilities" Feb 15 21:22:40 crc kubenswrapper[4735]: E0215 21:22:40.678722 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d6d340-66ed-43cc-9971-bee82135081d" containerName="registry-server" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.678728 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d6d340-66ed-43cc-9971-bee82135081d" containerName="registry-server" Feb 15 21:22:40 crc kubenswrapper[4735]: E0215 21:22:40.678737 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d6d340-66ed-43cc-9971-bee82135081d" containerName="extract-utilities" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.678744 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d6d340-66ed-43cc-9971-bee82135081d" containerName="extract-utilities" Feb 15 21:22:40 crc kubenswrapper[4735]: E0215 21:22:40.678754 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerName="extract-content" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.678759 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerName="extract-content" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.678969 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f380e363-4c49-4dfd-ab68-214b4c87015f" containerName="registry-server" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.678983 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d6d340-66ed-43cc-9971-bee82135081d" containerName="registry-server" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.680184 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.707780 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l92ph"] Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.810145 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-utilities\") pod \"redhat-operators-l92ph\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.810202 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zftts\" (UniqueName: \"kubernetes.io/projected/4a2204ae-8bcb-4100-8d5f-be78790d9df2-kube-api-access-zftts\") pod \"redhat-operators-l92ph\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.810612 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-catalog-content\") pod \"redhat-operators-l92ph\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.888477 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:22:40 crc kubenswrapper[4735]: E0215 21:22:40.889541 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.912242 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-utilities\") pod \"redhat-operators-l92ph\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.912304 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zftts\" (UniqueName: \"kubernetes.io/projected/4a2204ae-8bcb-4100-8d5f-be78790d9df2-kube-api-access-zftts\") pod \"redhat-operators-l92ph\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.912407 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-catalog-content\") pod \"redhat-operators-l92ph\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.912869 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-catalog-content\") pod \"redhat-operators-l92ph\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.913097 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-utilities\") pod \"redhat-operators-l92ph\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:40 crc kubenswrapper[4735]: I0215 21:22:40.937820 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zftts\" (UniqueName: \"kubernetes.io/projected/4a2204ae-8bcb-4100-8d5f-be78790d9df2-kube-api-access-zftts\") pod \"redhat-operators-l92ph\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:41 crc kubenswrapper[4735]: I0215 21:22:41.026011 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:41 crc kubenswrapper[4735]: I0215 21:22:41.513506 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l92ph"] Feb 15 21:22:41 crc kubenswrapper[4735]: I0215 21:22:41.856464 4735 generic.go:334] "Generic (PLEG): container finished" podID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerID="25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528" exitCode=0 Feb 15 21:22:41 crc kubenswrapper[4735]: I0215 21:22:41.856564 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92ph" event={"ID":"4a2204ae-8bcb-4100-8d5f-be78790d9df2","Type":"ContainerDied","Data":"25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528"} Feb 15 21:22:41 crc kubenswrapper[4735]: I0215 21:22:41.856767 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92ph" event={"ID":"4a2204ae-8bcb-4100-8d5f-be78790d9df2","Type":"ContainerStarted","Data":"bf5af565fc6b7f9f050a282a218a6a8799590e22778c549c1d9d5bd4e05bdcd6"} Feb 15 21:22:42 crc kubenswrapper[4735]: I0215 21:22:42.866672 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92ph" event={"ID":"4a2204ae-8bcb-4100-8d5f-be78790d9df2","Type":"ContainerStarted","Data":"a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be"} Feb 15 21:22:47 crc kubenswrapper[4735]: I0215 21:22:47.919592 4735 generic.go:334] "Generic (PLEG): container finished" podID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerID="a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be" exitCode=0 Feb 15 21:22:47 crc kubenswrapper[4735]: I0215 21:22:47.919672 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92ph" event={"ID":"4a2204ae-8bcb-4100-8d5f-be78790d9df2","Type":"ContainerDied","Data":"a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be"} Feb 15 21:22:48 crc kubenswrapper[4735]: I0215 21:22:48.930214 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92ph" event={"ID":"4a2204ae-8bcb-4100-8d5f-be78790d9df2","Type":"ContainerStarted","Data":"14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998"} Feb 15 21:22:51 crc kubenswrapper[4735]: I0215 21:22:51.026850 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:51 crc kubenswrapper[4735]: I0215 21:22:51.027224 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:22:52 crc kubenswrapper[4735]: I0215 21:22:52.074159 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l92ph" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="registry-server" probeResult="failure" output=< Feb 15 21:22:52 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:22:52 crc kubenswrapper[4735]: > Feb 15 21:22:53 crc kubenswrapper[4735]: I0215 21:22:53.887176 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:22:53 crc kubenswrapper[4735]: E0215 21:22:53.887779 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:23:02 crc kubenswrapper[4735]: I0215 21:23:02.099979 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l92ph" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="registry-server" probeResult="failure" output=< Feb 15 21:23:02 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:23:02 crc kubenswrapper[4735]: > Feb 15 21:23:04 crc kubenswrapper[4735]: I0215 21:23:04.891072 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:23:04 crc kubenswrapper[4735]: E0215 21:23:04.891708 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:23:11 crc kubenswrapper[4735]: I0215 21:23:11.092404 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:23:11 crc kubenswrapper[4735]: I0215 21:23:11.118054 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l92ph" podStartSLOduration=24.654004185 podStartE2EDuration="31.118029638s" podCreationTimestamp="2026-02-15 21:22:40 +0000 UTC" firstStartedPulling="2026-02-15 21:22:41.858628629 +0000 UTC m=+3969.724644252" lastFinishedPulling="2026-02-15 21:22:48.322654092 +0000 UTC m=+3976.188669705" observedRunningTime="2026-02-15 21:22:48.946499585 +0000 UTC m=+3976.812515208" watchObservedRunningTime="2026-02-15 21:23:11.118029638 +0000 UTC m=+3998.984045271" Feb 15 21:23:11 crc kubenswrapper[4735]: I0215 21:23:11.647845 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:23:11 crc kubenswrapper[4735]: I0215 21:23:11.884593 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l92ph"] Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.127730 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l92ph" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="registry-server" containerID="cri-o://14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998" gracePeriod=2 Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.789086 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.860555 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-utilities\") pod \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.860637 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zftts\" (UniqueName: \"kubernetes.io/projected/4a2204ae-8bcb-4100-8d5f-be78790d9df2-kube-api-access-zftts\") pod \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.860935 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-catalog-content\") pod \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\" (UID: \"4a2204ae-8bcb-4100-8d5f-be78790d9df2\") " Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.861529 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-utilities" (OuterVolumeSpecName: "utilities") pod "4a2204ae-8bcb-4100-8d5f-be78790d9df2" (UID: "4a2204ae-8bcb-4100-8d5f-be78790d9df2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.867013 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a2204ae-8bcb-4100-8d5f-be78790d9df2-kube-api-access-zftts" (OuterVolumeSpecName: "kube-api-access-zftts") pod "4a2204ae-8bcb-4100-8d5f-be78790d9df2" (UID: "4a2204ae-8bcb-4100-8d5f-be78790d9df2"). InnerVolumeSpecName "kube-api-access-zftts". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.963352 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zftts\" (UniqueName: \"kubernetes.io/projected/4a2204ae-8bcb-4100-8d5f-be78790d9df2-kube-api-access-zftts\") on node \"crc\" DevicePath \"\"" Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.963688 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:23:12 crc kubenswrapper[4735]: I0215 21:23:12.989169 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a2204ae-8bcb-4100-8d5f-be78790d9df2" (UID: "4a2204ae-8bcb-4100-8d5f-be78790d9df2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.065119 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a2204ae-8bcb-4100-8d5f-be78790d9df2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.136890 4735 generic.go:334] "Generic (PLEG): container finished" podID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerID="14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998" exitCode=0 Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.137035 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92ph" event={"ID":"4a2204ae-8bcb-4100-8d5f-be78790d9df2","Type":"ContainerDied","Data":"14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998"} Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.137102 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92ph" event={"ID":"4a2204ae-8bcb-4100-8d5f-be78790d9df2","Type":"ContainerDied","Data":"bf5af565fc6b7f9f050a282a218a6a8799590e22778c549c1d9d5bd4e05bdcd6"} Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.137143 4735 scope.go:117] "RemoveContainer" containerID="14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.138666 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l92ph" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.169174 4735 scope.go:117] "RemoveContainer" containerID="a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.188059 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l92ph"] Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.195470 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l92ph"] Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.214609 4735 scope.go:117] "RemoveContainer" containerID="25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.279265 4735 scope.go:117] "RemoveContainer" containerID="14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998" Feb 15 21:23:13 crc kubenswrapper[4735]: E0215 21:23:13.281552 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998\": container with ID starting with 14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998 not found: ID does not exist" containerID="14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.281624 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998"} err="failed to get container status \"14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998\": rpc error: code = NotFound desc = could not find container \"14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998\": container with ID starting with 14cf02280b5d29fca31e2fcf033b7a2df2441f3d9bac832e059ce98050c81998 not found: ID does not exist" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.281667 4735 scope.go:117] "RemoveContainer" containerID="a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be" Feb 15 21:23:13 crc kubenswrapper[4735]: E0215 21:23:13.282144 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be\": container with ID starting with a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be not found: ID does not exist" containerID="a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.282205 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be"} err="failed to get container status \"a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be\": rpc error: code = NotFound desc = could not find container \"a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be\": container with ID starting with a3bfa9382ef4b8a1e0bef255dde6788d2ea9fe39c2ab3128a8e808c30b3154be not found: ID does not exist" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.282240 4735 scope.go:117] "RemoveContainer" containerID="25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528" Feb 15 21:23:13 crc kubenswrapper[4735]: E0215 21:23:13.282509 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528\": container with ID starting with 25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528 not found: ID does not exist" containerID="25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528" Feb 15 21:23:13 crc kubenswrapper[4735]: I0215 21:23:13.282550 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528"} err="failed to get container status \"25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528\": rpc error: code = NotFound desc = could not find container \"25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528\": container with ID starting with 25428479b813c992055d2b107bdc9a90a9d781b9318a3e8767bbfcc3e8f9d528 not found: ID does not exist" Feb 15 21:23:14 crc kubenswrapper[4735]: I0215 21:23:14.898101 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" path="/var/lib/kubelet/pods/4a2204ae-8bcb-4100-8d5f-be78790d9df2/volumes" Feb 15 21:23:17 crc kubenswrapper[4735]: I0215 21:23:17.886984 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:23:17 crc kubenswrapper[4735]: E0215 21:23:17.887443 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:23:30 crc kubenswrapper[4735]: I0215 21:23:30.890313 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:23:30 crc kubenswrapper[4735]: E0215 21:23:30.891200 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:23:43 crc kubenswrapper[4735]: I0215 21:23:43.888077 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:23:43 crc kubenswrapper[4735]: E0215 21:23:43.889436 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:23:56 crc kubenswrapper[4735]: I0215 21:23:56.887722 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:23:57 crc kubenswrapper[4735]: I0215 21:23:57.564425 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"5402e50ecca53aef653cfbc5bd98553d9acbbcbcfa2fcbeb5b89235c75b76baf"} Feb 15 21:26:19 crc kubenswrapper[4735]: I0215 21:26:19.680170 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:26:19 crc kubenswrapper[4735]: I0215 21:26:19.680633 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.145176 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9qkrk"] Feb 15 21:26:21 crc kubenswrapper[4735]: E0215 21:26:21.146324 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="registry-server" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.146407 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="registry-server" Feb 15 21:26:21 crc kubenswrapper[4735]: E0215 21:26:21.146473 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="extract-content" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.146535 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="extract-content" Feb 15 21:26:21 crc kubenswrapper[4735]: E0215 21:26:21.146606 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="extract-utilities" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.146672 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="extract-utilities" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.146920 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a2204ae-8bcb-4100-8d5f-be78790d9df2" containerName="registry-server" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.149163 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.171556 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qkrk"] Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.318600 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvzj7\" (UniqueName: \"kubernetes.io/projected/141088a0-bf54-4845-aad3-dad181a1ed25-kube-api-access-xvzj7\") pod \"redhat-marketplace-9qkrk\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.319016 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-utilities\") pod \"redhat-marketplace-9qkrk\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.319140 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-catalog-content\") pod \"redhat-marketplace-9qkrk\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.421487 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvzj7\" (UniqueName: \"kubernetes.io/projected/141088a0-bf54-4845-aad3-dad181a1ed25-kube-api-access-xvzj7\") pod \"redhat-marketplace-9qkrk\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.421565 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-utilities\") pod \"redhat-marketplace-9qkrk\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.421629 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-catalog-content\") pod \"redhat-marketplace-9qkrk\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.422276 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-catalog-content\") pod \"redhat-marketplace-9qkrk\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.422285 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-utilities\") pod \"redhat-marketplace-9qkrk\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.454266 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvzj7\" (UniqueName: \"kubernetes.io/projected/141088a0-bf54-4845-aad3-dad181a1ed25-kube-api-access-xvzj7\") pod \"redhat-marketplace-9qkrk\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:21 crc kubenswrapper[4735]: I0215 21:26:21.468747 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:22 crc kubenswrapper[4735]: I0215 21:26:22.018076 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qkrk"] Feb 15 21:26:23 crc kubenswrapper[4735]: I0215 21:26:23.015770 4735 generic.go:334] "Generic (PLEG): container finished" podID="141088a0-bf54-4845-aad3-dad181a1ed25" containerID="9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d" exitCode=0 Feb 15 21:26:23 crc kubenswrapper[4735]: I0215 21:26:23.015878 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qkrk" event={"ID":"141088a0-bf54-4845-aad3-dad181a1ed25","Type":"ContainerDied","Data":"9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d"} Feb 15 21:26:23 crc kubenswrapper[4735]: I0215 21:26:23.016161 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qkrk" event={"ID":"141088a0-bf54-4845-aad3-dad181a1ed25","Type":"ContainerStarted","Data":"f03a8925411df2ab041fcaab85826492206031d50e0a8c294a2ded05e1324ec3"} Feb 15 21:26:24 crc kubenswrapper[4735]: I0215 21:26:24.026939 4735 generic.go:334] "Generic (PLEG): container finished" podID="abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" containerID="89280b91de1b8c16643f8e09110b4c71cba2e2e98b4c24d80ae67cc5a38fde2a" exitCode=0 Feb 15 21:26:24 crc kubenswrapper[4735]: I0215 21:26:24.027044 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383","Type":"ContainerDied","Data":"89280b91de1b8c16643f8e09110b4c71cba2e2e98b4c24d80ae67cc5a38fde2a"} Feb 15 21:26:24 crc kubenswrapper[4735]: I0215 21:26:24.030610 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qkrk" event={"ID":"141088a0-bf54-4845-aad3-dad181a1ed25","Type":"ContainerStarted","Data":"bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb"} Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.038059 4735 generic.go:334] "Generic (PLEG): container finished" podID="141088a0-bf54-4845-aad3-dad181a1ed25" containerID="bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb" exitCode=0 Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.039255 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qkrk" event={"ID":"141088a0-bf54-4845-aad3-dad181a1ed25","Type":"ContainerDied","Data":"bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb"} Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.501507 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.625760 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psbch\" (UniqueName: \"kubernetes.io/projected/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-kube-api-access-psbch\") pod \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.625823 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-config-data\") pod \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.625852 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-temporary\") pod \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.626005 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ca-certs\") pod \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.626103 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config-secret\") pod \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.626139 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-workdir\") pod \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.626195 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.626249 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config\") pod \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.626324 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ssh-key\") pod \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\" (UID: \"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383\") " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.628759 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" (UID: "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.630554 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-config-data" (OuterVolumeSpecName: "config-data") pod "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" (UID: "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.634843 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" (UID: "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.645080 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" (UID: "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.646171 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-kube-api-access-psbch" (OuterVolumeSpecName: "kube-api-access-psbch") pod "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" (UID: "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383"). InnerVolumeSpecName "kube-api-access-psbch". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.666721 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" (UID: "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.667532 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" (UID: "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.672856 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" (UID: "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.718414 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" (UID: "abcf2b9d-8410-4a1b-be0e-3dde2a6e3383"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.728502 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psbch\" (UniqueName: \"kubernetes.io/projected/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-kube-api-access-psbch\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.728923 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-config-data\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.728956 4735 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.728966 4735 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.728975 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.728984 4735 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.729443 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.729468 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.729484 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abcf2b9d-8410-4a1b-be0e-3dde2a6e3383-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.768740 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 15 21:26:25 crc kubenswrapper[4735]: I0215 21:26:25.836247 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:26 crc kubenswrapper[4735]: I0215 21:26:26.048373 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qkrk" event={"ID":"141088a0-bf54-4845-aad3-dad181a1ed25","Type":"ContainerStarted","Data":"ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121"} Feb 15 21:26:26 crc kubenswrapper[4735]: I0215 21:26:26.051172 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"abcf2b9d-8410-4a1b-be0e-3dde2a6e3383","Type":"ContainerDied","Data":"5a4e6b489f4c322458d9ff0b86f8a712a576a93e47188959ad07a2cb7f80f137"} Feb 15 21:26:26 crc kubenswrapper[4735]: I0215 21:26:26.051199 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a4e6b489f4c322458d9ff0b86f8a712a576a93e47188959ad07a2cb7f80f137" Feb 15 21:26:26 crc kubenswrapper[4735]: I0215 21:26:26.051261 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 15 21:26:26 crc kubenswrapper[4735]: I0215 21:26:26.090403 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9qkrk" podStartSLOduration=2.653581112 podStartE2EDuration="5.09038583s" podCreationTimestamp="2026-02-15 21:26:21 +0000 UTC" firstStartedPulling="2026-02-15 21:26:23.017557153 +0000 UTC m=+4190.883572796" lastFinishedPulling="2026-02-15 21:26:25.454361891 +0000 UTC m=+4193.320377514" observedRunningTime="2026-02-15 21:26:26.086317712 +0000 UTC m=+4193.952333335" watchObservedRunningTime="2026-02-15 21:26:26.09038583 +0000 UTC m=+4193.956401453" Feb 15 21:26:31 crc kubenswrapper[4735]: I0215 21:26:31.469514 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:31 crc kubenswrapper[4735]: I0215 21:26:31.471158 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:31 crc kubenswrapper[4735]: I0215 21:26:31.524430 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:32 crc kubenswrapper[4735]: I0215 21:26:32.178638 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:32 crc kubenswrapper[4735]: I0215 21:26:32.249451 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qkrk"] Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.133133 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9qkrk" podUID="141088a0-bf54-4845-aad3-dad181a1ed25" containerName="registry-server" containerID="cri-o://ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121" gracePeriod=2 Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.314125 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 15 21:26:34 crc kubenswrapper[4735]: E0215 21:26:34.314732 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" containerName="tempest-tests-tempest-tests-runner" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.314744 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" containerName="tempest-tests-tempest-tests-runner" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.314955 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="abcf2b9d-8410-4a1b-be0e-3dde2a6e3383" containerName="tempest-tests-tempest-tests-runner" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.316250 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.320517 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-j75vs" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.349343 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.433213 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"97a6ddef-7ccf-4c1f-bc90-691f4511e4ad\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.433413 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zps7p\" (UniqueName: \"kubernetes.io/projected/97a6ddef-7ccf-4c1f-bc90-691f4511e4ad-kube-api-access-zps7p\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"97a6ddef-7ccf-4c1f-bc90-691f4511e4ad\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.535442 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"97a6ddef-7ccf-4c1f-bc90-691f4511e4ad\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.535624 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zps7p\" (UniqueName: \"kubernetes.io/projected/97a6ddef-7ccf-4c1f-bc90-691f4511e4ad-kube-api-access-zps7p\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"97a6ddef-7ccf-4c1f-bc90-691f4511e4ad\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.541236 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"97a6ddef-7ccf-4c1f-bc90-691f4511e4ad\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.582020 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zps7p\" (UniqueName: \"kubernetes.io/projected/97a6ddef-7ccf-4c1f-bc90-691f4511e4ad-kube-api-access-zps7p\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"97a6ddef-7ccf-4c1f-bc90-691f4511e4ad\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.604017 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"97a6ddef-7ccf-4c1f-bc90-691f4511e4ad\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.645446 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.771620 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.942473 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-utilities\") pod \"141088a0-bf54-4845-aad3-dad181a1ed25\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.942640 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvzj7\" (UniqueName: \"kubernetes.io/projected/141088a0-bf54-4845-aad3-dad181a1ed25-kube-api-access-xvzj7\") pod \"141088a0-bf54-4845-aad3-dad181a1ed25\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.942704 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-catalog-content\") pod \"141088a0-bf54-4845-aad3-dad181a1ed25\" (UID: \"141088a0-bf54-4845-aad3-dad181a1ed25\") " Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.943395 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-utilities" (OuterVolumeSpecName: "utilities") pod "141088a0-bf54-4845-aad3-dad181a1ed25" (UID: "141088a0-bf54-4845-aad3-dad181a1ed25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.946886 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/141088a0-bf54-4845-aad3-dad181a1ed25-kube-api-access-xvzj7" (OuterVolumeSpecName: "kube-api-access-xvzj7") pod "141088a0-bf54-4845-aad3-dad181a1ed25" (UID: "141088a0-bf54-4845-aad3-dad181a1ed25"). InnerVolumeSpecName "kube-api-access-xvzj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:26:34 crc kubenswrapper[4735]: I0215 21:26:34.971755 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "141088a0-bf54-4845-aad3-dad181a1ed25" (UID: "141088a0-bf54-4845-aad3-dad181a1ed25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.045506 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.045542 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvzj7\" (UniqueName: \"kubernetes.io/projected/141088a0-bf54-4845-aad3-dad181a1ed25-kube-api-access-xvzj7\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.045552 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/141088a0-bf54-4845-aad3-dad181a1ed25-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.106298 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.141225 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"97a6ddef-7ccf-4c1f-bc90-691f4511e4ad","Type":"ContainerStarted","Data":"ba19747141edbd8aae63ab2c02f2e2cc263ce6ddde49c9a36d3f48810823f593"} Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.145087 4735 generic.go:334] "Generic (PLEG): container finished" podID="141088a0-bf54-4845-aad3-dad181a1ed25" containerID="ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121" exitCode=0 Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.145132 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qkrk" event={"ID":"141088a0-bf54-4845-aad3-dad181a1ed25","Type":"ContainerDied","Data":"ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121"} Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.145163 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qkrk" event={"ID":"141088a0-bf54-4845-aad3-dad181a1ed25","Type":"ContainerDied","Data":"f03a8925411df2ab041fcaab85826492206031d50e0a8c294a2ded05e1324ec3"} Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.145182 4735 scope.go:117] "RemoveContainer" containerID="ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.145295 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qkrk" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.163678 4735 scope.go:117] "RemoveContainer" containerID="bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.194152 4735 scope.go:117] "RemoveContainer" containerID="9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.198102 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qkrk"] Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.210756 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qkrk"] Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.215417 4735 scope.go:117] "RemoveContainer" containerID="ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121" Feb 15 21:26:35 crc kubenswrapper[4735]: E0215 21:26:35.215764 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121\": container with ID starting with ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121 not found: ID does not exist" containerID="ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.215797 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121"} err="failed to get container status \"ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121\": rpc error: code = NotFound desc = could not find container \"ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121\": container with ID starting with ce43b7771f5ce183060b484555cbe8b4cfe41b4d94fb3574e2f7e4ad4be6e121 not found: ID does not exist" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.215817 4735 scope.go:117] "RemoveContainer" containerID="bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb" Feb 15 21:26:35 crc kubenswrapper[4735]: E0215 21:26:35.216715 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb\": container with ID starting with bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb not found: ID does not exist" containerID="bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.216771 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb"} err="failed to get container status \"bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb\": rpc error: code = NotFound desc = could not find container \"bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb\": container with ID starting with bf2510a60a79a105f3826fd09d082154e33609f88ac16774f59ef41dba7564cb not found: ID does not exist" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.216907 4735 scope.go:117] "RemoveContainer" containerID="9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d" Feb 15 21:26:35 crc kubenswrapper[4735]: E0215 21:26:35.217354 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d\": container with ID starting with 9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d not found: ID does not exist" containerID="9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d" Feb 15 21:26:35 crc kubenswrapper[4735]: I0215 21:26:35.217389 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d"} err="failed to get container status \"9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d\": rpc error: code = NotFound desc = could not find container \"9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d\": container with ID starting with 9da62ddaa697f00b2b6e6cea7cd90dd35227edf75a81c5a240ea69ec49bb4f0d not found: ID does not exist" Feb 15 21:26:36 crc kubenswrapper[4735]: I0215 21:26:36.905395 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="141088a0-bf54-4845-aad3-dad181a1ed25" path="/var/lib/kubelet/pods/141088a0-bf54-4845-aad3-dad181a1ed25/volumes" Feb 15 21:26:37 crc kubenswrapper[4735]: I0215 21:26:37.168343 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"97a6ddef-7ccf-4c1f-bc90-691f4511e4ad","Type":"ContainerStarted","Data":"c0a848887d1231d8c74e9627e56f4bd157fd54a0351f4da82629aa5882273275"} Feb 15 21:26:37 crc kubenswrapper[4735]: I0215 21:26:37.203564 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.041163697 podStartE2EDuration="3.203534911s" podCreationTimestamp="2026-02-15 21:26:34 +0000 UTC" firstStartedPulling="2026-02-15 21:26:35.117141384 +0000 UTC m=+4202.983157007" lastFinishedPulling="2026-02-15 21:26:36.279512588 +0000 UTC m=+4204.145528221" observedRunningTime="2026-02-15 21:26:37.193928104 +0000 UTC m=+4205.059943767" watchObservedRunningTime="2026-02-15 21:26:37.203534911 +0000 UTC m=+4205.069550564" Feb 15 21:26:49 crc kubenswrapper[4735]: I0215 21:26:49.679718 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:26:49 crc kubenswrapper[4735]: I0215 21:26:49.680208 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.114040 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4fgc7/must-gather-b78nv"] Feb 15 21:26:58 crc kubenswrapper[4735]: E0215 21:26:58.114922 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="141088a0-bf54-4845-aad3-dad181a1ed25" containerName="registry-server" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.114934 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="141088a0-bf54-4845-aad3-dad181a1ed25" containerName="registry-server" Feb 15 21:26:58 crc kubenswrapper[4735]: E0215 21:26:58.114961 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="141088a0-bf54-4845-aad3-dad181a1ed25" containerName="extract-content" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.114968 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="141088a0-bf54-4845-aad3-dad181a1ed25" containerName="extract-content" Feb 15 21:26:58 crc kubenswrapper[4735]: E0215 21:26:58.114997 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="141088a0-bf54-4845-aad3-dad181a1ed25" containerName="extract-utilities" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.115005 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="141088a0-bf54-4845-aad3-dad181a1ed25" containerName="extract-utilities" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.115166 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="141088a0-bf54-4845-aad3-dad181a1ed25" containerName="registry-server" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.116039 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.120653 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4fgc7"/"kube-root-ca.crt" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.120809 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4fgc7"/"openshift-service-ca.crt" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.123733 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4fgc7"/"default-dockercfg-pkmn8" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.139612 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4fgc7/must-gather-b78nv"] Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.207482 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/694be1b6-7203-4035-a89f-e254957c6311-must-gather-output\") pod \"must-gather-b78nv\" (UID: \"694be1b6-7203-4035-a89f-e254957c6311\") " pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.208363 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rt2c\" (UniqueName: \"kubernetes.io/projected/694be1b6-7203-4035-a89f-e254957c6311-kube-api-access-2rt2c\") pod \"must-gather-b78nv\" (UID: \"694be1b6-7203-4035-a89f-e254957c6311\") " pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.310208 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rt2c\" (UniqueName: \"kubernetes.io/projected/694be1b6-7203-4035-a89f-e254957c6311-kube-api-access-2rt2c\") pod \"must-gather-b78nv\" (UID: \"694be1b6-7203-4035-a89f-e254957c6311\") " pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.310590 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/694be1b6-7203-4035-a89f-e254957c6311-must-gather-output\") pod \"must-gather-b78nv\" (UID: \"694be1b6-7203-4035-a89f-e254957c6311\") " pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.310887 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/694be1b6-7203-4035-a89f-e254957c6311-must-gather-output\") pod \"must-gather-b78nv\" (UID: \"694be1b6-7203-4035-a89f-e254957c6311\") " pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.339508 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rt2c\" (UniqueName: \"kubernetes.io/projected/694be1b6-7203-4035-a89f-e254957c6311-kube-api-access-2rt2c\") pod \"must-gather-b78nv\" (UID: \"694be1b6-7203-4035-a89f-e254957c6311\") " pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.435093 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.973759 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4fgc7/must-gather-b78nv"] Feb 15 21:26:58 crc kubenswrapper[4735]: I0215 21:26:58.985076 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 21:26:59 crc kubenswrapper[4735]: I0215 21:26:59.385321 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/must-gather-b78nv" event={"ID":"694be1b6-7203-4035-a89f-e254957c6311","Type":"ContainerStarted","Data":"d769e86f00a61cd7e2593eb0066edba5bc5b1f36ef261b7609c1f6042c9e7ab5"} Feb 15 21:27:06 crc kubenswrapper[4735]: I0215 21:27:06.446618 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/must-gather-b78nv" event={"ID":"694be1b6-7203-4035-a89f-e254957c6311","Type":"ContainerStarted","Data":"9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991"} Feb 15 21:27:07 crc kubenswrapper[4735]: I0215 21:27:07.456571 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/must-gather-b78nv" event={"ID":"694be1b6-7203-4035-a89f-e254957c6311","Type":"ContainerStarted","Data":"bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734"} Feb 15 21:27:07 crc kubenswrapper[4735]: I0215 21:27:07.476894 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4fgc7/must-gather-b78nv" podStartSLOduration=2.366556118 podStartE2EDuration="9.476877619s" podCreationTimestamp="2026-02-15 21:26:58 +0000 UTC" firstStartedPulling="2026-02-15 21:26:58.984868285 +0000 UTC m=+4226.850883908" lastFinishedPulling="2026-02-15 21:27:06.095189786 +0000 UTC m=+4233.961205409" observedRunningTime="2026-02-15 21:27:07.472493851 +0000 UTC m=+4235.338509474" watchObservedRunningTime="2026-02-15 21:27:07.476877619 +0000 UTC m=+4235.342893242" Feb 15 21:27:11 crc kubenswrapper[4735]: I0215 21:27:11.433329 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4fgc7/crc-debug-vwnvz"] Feb 15 21:27:11 crc kubenswrapper[4735]: I0215 21:27:11.434892 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:27:11 crc kubenswrapper[4735]: I0215 21:27:11.505599 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/447c478e-86b9-4af3-9766-be0cb4b83873-host\") pod \"crc-debug-vwnvz\" (UID: \"447c478e-86b9-4af3-9766-be0cb4b83873\") " pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:27:11 crc kubenswrapper[4735]: I0215 21:27:11.506063 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnqgc\" (UniqueName: \"kubernetes.io/projected/447c478e-86b9-4af3-9766-be0cb4b83873-kube-api-access-xnqgc\") pod \"crc-debug-vwnvz\" (UID: \"447c478e-86b9-4af3-9766-be0cb4b83873\") " pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:27:11 crc kubenswrapper[4735]: I0215 21:27:11.607546 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/447c478e-86b9-4af3-9766-be0cb4b83873-host\") pod \"crc-debug-vwnvz\" (UID: \"447c478e-86b9-4af3-9766-be0cb4b83873\") " pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:27:11 crc kubenswrapper[4735]: I0215 21:27:11.607652 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnqgc\" (UniqueName: \"kubernetes.io/projected/447c478e-86b9-4af3-9766-be0cb4b83873-kube-api-access-xnqgc\") pod \"crc-debug-vwnvz\" (UID: \"447c478e-86b9-4af3-9766-be0cb4b83873\") " pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:27:11 crc kubenswrapper[4735]: I0215 21:27:11.608566 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/447c478e-86b9-4af3-9766-be0cb4b83873-host\") pod \"crc-debug-vwnvz\" (UID: \"447c478e-86b9-4af3-9766-be0cb4b83873\") " pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:27:11 crc kubenswrapper[4735]: I0215 21:27:11.629682 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnqgc\" (UniqueName: \"kubernetes.io/projected/447c478e-86b9-4af3-9766-be0cb4b83873-kube-api-access-xnqgc\") pod \"crc-debug-vwnvz\" (UID: \"447c478e-86b9-4af3-9766-be0cb4b83873\") " pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:27:11 crc kubenswrapper[4735]: I0215 21:27:11.751685 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:27:12 crc kubenswrapper[4735]: I0215 21:27:12.512051 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" event={"ID":"447c478e-86b9-4af3-9766-be0cb4b83873","Type":"ContainerStarted","Data":"21458832a23c2949174cb986a3073a819635292a66bfb0ac43a73efee02b49c1"} Feb 15 21:27:19 crc kubenswrapper[4735]: I0215 21:27:19.679931 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:27:19 crc kubenswrapper[4735]: I0215 21:27:19.680472 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:27:19 crc kubenswrapper[4735]: I0215 21:27:19.680522 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 21:27:19 crc kubenswrapper[4735]: I0215 21:27:19.681229 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5402e50ecca53aef653cfbc5bd98553d9acbbcbcfa2fcbeb5b89235c75b76baf"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 21:27:19 crc kubenswrapper[4735]: I0215 21:27:19.681277 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://5402e50ecca53aef653cfbc5bd98553d9acbbcbcfa2fcbeb5b89235c75b76baf" gracePeriod=600 Feb 15 21:27:20 crc kubenswrapper[4735]: I0215 21:27:20.595531 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="5402e50ecca53aef653cfbc5bd98553d9acbbcbcfa2fcbeb5b89235c75b76baf" exitCode=0 Feb 15 21:27:20 crc kubenswrapper[4735]: I0215 21:27:20.595595 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"5402e50ecca53aef653cfbc5bd98553d9acbbcbcfa2fcbeb5b89235c75b76baf"} Feb 15 21:27:20 crc kubenswrapper[4735]: I0215 21:27:20.595849 4735 scope.go:117] "RemoveContainer" containerID="5f8068485387f714132c02c0fe4ed7d2591ceee110c06b262489b337dc9127ca" Feb 15 21:27:22 crc kubenswrapper[4735]: I0215 21:27:22.623672 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91"} Feb 15 21:27:22 crc kubenswrapper[4735]: I0215 21:27:22.626467 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" event={"ID":"447c478e-86b9-4af3-9766-be0cb4b83873","Type":"ContainerStarted","Data":"d963cf4e01a2739498b2f98fb02b10f4dc5e09e72b52c7e6f1a1c7949ff0f3c0"} Feb 15 21:27:22 crc kubenswrapper[4735]: I0215 21:27:22.656291 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" podStartSLOduration=1.922960147 podStartE2EDuration="11.656271804s" podCreationTimestamp="2026-02-15 21:27:11 +0000 UTC" firstStartedPulling="2026-02-15 21:27:11.784853865 +0000 UTC m=+4239.650869488" lastFinishedPulling="2026-02-15 21:27:21.518165522 +0000 UTC m=+4249.384181145" observedRunningTime="2026-02-15 21:27:22.652822322 +0000 UTC m=+4250.518837945" watchObservedRunningTime="2026-02-15 21:27:22.656271804 +0000 UTC m=+4250.522287427" Feb 15 21:28:08 crc kubenswrapper[4735]: I0215 21:28:08.026507 4735 generic.go:334] "Generic (PLEG): container finished" podID="447c478e-86b9-4af3-9766-be0cb4b83873" containerID="d963cf4e01a2739498b2f98fb02b10f4dc5e09e72b52c7e6f1a1c7949ff0f3c0" exitCode=0 Feb 15 21:28:08 crc kubenswrapper[4735]: I0215 21:28:08.026678 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" event={"ID":"447c478e-86b9-4af3-9766-be0cb4b83873","Type":"ContainerDied","Data":"d963cf4e01a2739498b2f98fb02b10f4dc5e09e72b52c7e6f1a1c7949ff0f3c0"} Feb 15 21:28:09 crc kubenswrapper[4735]: I0215 21:28:09.142033 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:28:09 crc kubenswrapper[4735]: I0215 21:28:09.176248 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4fgc7/crc-debug-vwnvz"] Feb 15 21:28:09 crc kubenswrapper[4735]: I0215 21:28:09.185016 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4fgc7/crc-debug-vwnvz"] Feb 15 21:28:09 crc kubenswrapper[4735]: I0215 21:28:09.270780 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnqgc\" (UniqueName: \"kubernetes.io/projected/447c478e-86b9-4af3-9766-be0cb4b83873-kube-api-access-xnqgc\") pod \"447c478e-86b9-4af3-9766-be0cb4b83873\" (UID: \"447c478e-86b9-4af3-9766-be0cb4b83873\") " Feb 15 21:28:09 crc kubenswrapper[4735]: I0215 21:28:09.271234 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/447c478e-86b9-4af3-9766-be0cb4b83873-host\") pod \"447c478e-86b9-4af3-9766-be0cb4b83873\" (UID: \"447c478e-86b9-4af3-9766-be0cb4b83873\") " Feb 15 21:28:09 crc kubenswrapper[4735]: I0215 21:28:09.271324 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/447c478e-86b9-4af3-9766-be0cb4b83873-host" (OuterVolumeSpecName: "host") pod "447c478e-86b9-4af3-9766-be0cb4b83873" (UID: "447c478e-86b9-4af3-9766-be0cb4b83873"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 21:28:09 crc kubenswrapper[4735]: I0215 21:28:09.271696 4735 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/447c478e-86b9-4af3-9766-be0cb4b83873-host\") on node \"crc\" DevicePath \"\"" Feb 15 21:28:09 crc kubenswrapper[4735]: I0215 21:28:09.285094 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/447c478e-86b9-4af3-9766-be0cb4b83873-kube-api-access-xnqgc" (OuterVolumeSpecName: "kube-api-access-xnqgc") pod "447c478e-86b9-4af3-9766-be0cb4b83873" (UID: "447c478e-86b9-4af3-9766-be0cb4b83873"). InnerVolumeSpecName "kube-api-access-xnqgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:28:09 crc kubenswrapper[4735]: I0215 21:28:09.373634 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnqgc\" (UniqueName: \"kubernetes.io/projected/447c478e-86b9-4af3-9766-be0cb4b83873-kube-api-access-xnqgc\") on node \"crc\" DevicePath \"\"" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.059908 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21458832a23c2949174cb986a3073a819635292a66bfb0ac43a73efee02b49c1" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.060098 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-vwnvz" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.363120 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4fgc7/crc-debug-t6jvl"] Feb 15 21:28:10 crc kubenswrapper[4735]: E0215 21:28:10.363750 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447c478e-86b9-4af3-9766-be0cb4b83873" containerName="container-00" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.363762 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="447c478e-86b9-4af3-9766-be0cb4b83873" containerName="container-00" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.364023 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="447c478e-86b9-4af3-9766-be0cb4b83873" containerName="container-00" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.364627 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.391750 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-host\") pod \"crc-debug-t6jvl\" (UID: \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\") " pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.391902 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnf2m\" (UniqueName: \"kubernetes.io/projected/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-kube-api-access-jnf2m\") pod \"crc-debug-t6jvl\" (UID: \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\") " pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.494186 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnf2m\" (UniqueName: \"kubernetes.io/projected/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-kube-api-access-jnf2m\") pod \"crc-debug-t6jvl\" (UID: \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\") " pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.494269 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-host\") pod \"crc-debug-t6jvl\" (UID: \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\") " pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.494373 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-host\") pod \"crc-debug-t6jvl\" (UID: \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\") " pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.517636 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnf2m\" (UniqueName: \"kubernetes.io/projected/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-kube-api-access-jnf2m\") pod \"crc-debug-t6jvl\" (UID: \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\") " pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.680671 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:10 crc kubenswrapper[4735]: I0215 21:28:10.895504 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="447c478e-86b9-4af3-9766-be0cb4b83873" path="/var/lib/kubelet/pods/447c478e-86b9-4af3-9766-be0cb4b83873/volumes" Feb 15 21:28:11 crc kubenswrapper[4735]: I0215 21:28:11.067920 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" event={"ID":"35f06bdb-3cb1-4821-bf30-e2d3d5df8695","Type":"ContainerStarted","Data":"07391ba50ab8f8b18b852f5e536b0e09724bfe790306f200ddd8bc8cef71ea58"} Feb 15 21:28:11 crc kubenswrapper[4735]: I0215 21:28:11.067983 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" event={"ID":"35f06bdb-3cb1-4821-bf30-e2d3d5df8695","Type":"ContainerStarted","Data":"61a11368e0ba5d41e3381e52b555d24478d99b5e43794d26c2f55a8bb78c765d"} Feb 15 21:28:11 crc kubenswrapper[4735]: I0215 21:28:11.088642 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" podStartSLOduration=1.08862577 podStartE2EDuration="1.08862577s" podCreationTimestamp="2026-02-15 21:28:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 21:28:11.084296684 +0000 UTC m=+4298.950312307" watchObservedRunningTime="2026-02-15 21:28:11.08862577 +0000 UTC m=+4298.954641393" Feb 15 21:28:12 crc kubenswrapper[4735]: I0215 21:28:12.077802 4735 generic.go:334] "Generic (PLEG): container finished" podID="35f06bdb-3cb1-4821-bf30-e2d3d5df8695" containerID="07391ba50ab8f8b18b852f5e536b0e09724bfe790306f200ddd8bc8cef71ea58" exitCode=0 Feb 15 21:28:12 crc kubenswrapper[4735]: I0215 21:28:12.078318 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" event={"ID":"35f06bdb-3cb1-4821-bf30-e2d3d5df8695","Type":"ContainerDied","Data":"07391ba50ab8f8b18b852f5e536b0e09724bfe790306f200ddd8bc8cef71ea58"} Feb 15 21:28:13 crc kubenswrapper[4735]: I0215 21:28:13.186399 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:13 crc kubenswrapper[4735]: I0215 21:28:13.218605 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4fgc7/crc-debug-t6jvl"] Feb 15 21:28:13 crc kubenswrapper[4735]: I0215 21:28:13.235298 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4fgc7/crc-debug-t6jvl"] Feb 15 21:28:13 crc kubenswrapper[4735]: I0215 21:28:13.345881 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-host\") pod \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\" (UID: \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\") " Feb 15 21:28:13 crc kubenswrapper[4735]: I0215 21:28:13.345981 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnf2m\" (UniqueName: \"kubernetes.io/projected/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-kube-api-access-jnf2m\") pod \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\" (UID: \"35f06bdb-3cb1-4821-bf30-e2d3d5df8695\") " Feb 15 21:28:13 crc kubenswrapper[4735]: I0215 21:28:13.346198 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-host" (OuterVolumeSpecName: "host") pod "35f06bdb-3cb1-4821-bf30-e2d3d5df8695" (UID: "35f06bdb-3cb1-4821-bf30-e2d3d5df8695"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 21:28:13 crc kubenswrapper[4735]: I0215 21:28:13.346668 4735 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-host\") on node \"crc\" DevicePath \"\"" Feb 15 21:28:13 crc kubenswrapper[4735]: I0215 21:28:13.360560 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-kube-api-access-jnf2m" (OuterVolumeSpecName: "kube-api-access-jnf2m") pod "35f06bdb-3cb1-4821-bf30-e2d3d5df8695" (UID: "35f06bdb-3cb1-4821-bf30-e2d3d5df8695"). InnerVolumeSpecName "kube-api-access-jnf2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:28:13 crc kubenswrapper[4735]: I0215 21:28:13.448617 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnf2m\" (UniqueName: \"kubernetes.io/projected/35f06bdb-3cb1-4821-bf30-e2d3d5df8695-kube-api-access-jnf2m\") on node \"crc\" DevicePath \"\"" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.100365 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61a11368e0ba5d41e3381e52b555d24478d99b5e43794d26c2f55a8bb78c765d" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.100410 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-t6jvl" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.508070 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4fgc7/crc-debug-kpm99"] Feb 15 21:28:14 crc kubenswrapper[4735]: E0215 21:28:14.508510 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f06bdb-3cb1-4821-bf30-e2d3d5df8695" containerName="container-00" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.508525 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f06bdb-3cb1-4821-bf30-e2d3d5df8695" containerName="container-00" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.508728 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="35f06bdb-3cb1-4821-bf30-e2d3d5df8695" containerName="container-00" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.509373 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.668602 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzbqb\" (UniqueName: \"kubernetes.io/projected/fc62a3ab-f35e-480f-b1aa-3320825d180b-kube-api-access-nzbqb\") pod \"crc-debug-kpm99\" (UID: \"fc62a3ab-f35e-480f-b1aa-3320825d180b\") " pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.668663 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc62a3ab-f35e-480f-b1aa-3320825d180b-host\") pod \"crc-debug-kpm99\" (UID: \"fc62a3ab-f35e-480f-b1aa-3320825d180b\") " pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.770746 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzbqb\" (UniqueName: \"kubernetes.io/projected/fc62a3ab-f35e-480f-b1aa-3320825d180b-kube-api-access-nzbqb\") pod \"crc-debug-kpm99\" (UID: \"fc62a3ab-f35e-480f-b1aa-3320825d180b\") " pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.771185 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc62a3ab-f35e-480f-b1aa-3320825d180b-host\") pod \"crc-debug-kpm99\" (UID: \"fc62a3ab-f35e-480f-b1aa-3320825d180b\") " pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.771280 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc62a3ab-f35e-480f-b1aa-3320825d180b-host\") pod \"crc-debug-kpm99\" (UID: \"fc62a3ab-f35e-480f-b1aa-3320825d180b\") " pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.795643 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzbqb\" (UniqueName: \"kubernetes.io/projected/fc62a3ab-f35e-480f-b1aa-3320825d180b-kube-api-access-nzbqb\") pod \"crc-debug-kpm99\" (UID: \"fc62a3ab-f35e-480f-b1aa-3320825d180b\") " pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.830231 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:14 crc kubenswrapper[4735]: W0215 21:28:14.854212 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc62a3ab_f35e_480f_b1aa_3320825d180b.slice/crio-e046805a5f68bcbf4c07361e04fb84fcd5cdbdff0b9594c49da3c65d0e727278 WatchSource:0}: Error finding container e046805a5f68bcbf4c07361e04fb84fcd5cdbdff0b9594c49da3c65d0e727278: Status 404 returned error can't find the container with id e046805a5f68bcbf4c07361e04fb84fcd5cdbdff0b9594c49da3c65d0e727278 Feb 15 21:28:14 crc kubenswrapper[4735]: I0215 21:28:14.902326 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35f06bdb-3cb1-4821-bf30-e2d3d5df8695" path="/var/lib/kubelet/pods/35f06bdb-3cb1-4821-bf30-e2d3d5df8695/volumes" Feb 15 21:28:15 crc kubenswrapper[4735]: I0215 21:28:15.111372 4735 generic.go:334] "Generic (PLEG): container finished" podID="fc62a3ab-f35e-480f-b1aa-3320825d180b" containerID="5da3e571ad86b50bbb55d7d565f8277689d6fc25fa6163541ac1e110cd8263d3" exitCode=0 Feb 15 21:28:15 crc kubenswrapper[4735]: I0215 21:28:15.111448 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/crc-debug-kpm99" event={"ID":"fc62a3ab-f35e-480f-b1aa-3320825d180b","Type":"ContainerDied","Data":"5da3e571ad86b50bbb55d7d565f8277689d6fc25fa6163541ac1e110cd8263d3"} Feb 15 21:28:15 crc kubenswrapper[4735]: I0215 21:28:15.111793 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/crc-debug-kpm99" event={"ID":"fc62a3ab-f35e-480f-b1aa-3320825d180b","Type":"ContainerStarted","Data":"e046805a5f68bcbf4c07361e04fb84fcd5cdbdff0b9594c49da3c65d0e727278"} Feb 15 21:28:15 crc kubenswrapper[4735]: I0215 21:28:15.152172 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4fgc7/crc-debug-kpm99"] Feb 15 21:28:15 crc kubenswrapper[4735]: I0215 21:28:15.160475 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4fgc7/crc-debug-kpm99"] Feb 15 21:28:16 crc kubenswrapper[4735]: I0215 21:28:16.219304 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:16 crc kubenswrapper[4735]: I0215 21:28:16.412837 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc62a3ab-f35e-480f-b1aa-3320825d180b-host\") pod \"fc62a3ab-f35e-480f-b1aa-3320825d180b\" (UID: \"fc62a3ab-f35e-480f-b1aa-3320825d180b\") " Feb 15 21:28:16 crc kubenswrapper[4735]: I0215 21:28:16.412935 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc62a3ab-f35e-480f-b1aa-3320825d180b-host" (OuterVolumeSpecName: "host") pod "fc62a3ab-f35e-480f-b1aa-3320825d180b" (UID: "fc62a3ab-f35e-480f-b1aa-3320825d180b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 21:28:16 crc kubenswrapper[4735]: I0215 21:28:16.413070 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzbqb\" (UniqueName: \"kubernetes.io/projected/fc62a3ab-f35e-480f-b1aa-3320825d180b-kube-api-access-nzbqb\") pod \"fc62a3ab-f35e-480f-b1aa-3320825d180b\" (UID: \"fc62a3ab-f35e-480f-b1aa-3320825d180b\") " Feb 15 21:28:16 crc kubenswrapper[4735]: I0215 21:28:16.413596 4735 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc62a3ab-f35e-480f-b1aa-3320825d180b-host\") on node \"crc\" DevicePath \"\"" Feb 15 21:28:16 crc kubenswrapper[4735]: I0215 21:28:16.423135 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc62a3ab-f35e-480f-b1aa-3320825d180b-kube-api-access-nzbqb" (OuterVolumeSpecName: "kube-api-access-nzbqb") pod "fc62a3ab-f35e-480f-b1aa-3320825d180b" (UID: "fc62a3ab-f35e-480f-b1aa-3320825d180b"). InnerVolumeSpecName "kube-api-access-nzbqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:28:16 crc kubenswrapper[4735]: I0215 21:28:16.516451 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzbqb\" (UniqueName: \"kubernetes.io/projected/fc62a3ab-f35e-480f-b1aa-3320825d180b-kube-api-access-nzbqb\") on node \"crc\" DevicePath \"\"" Feb 15 21:28:16 crc kubenswrapper[4735]: I0215 21:28:16.897562 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc62a3ab-f35e-480f-b1aa-3320825d180b" path="/var/lib/kubelet/pods/fc62a3ab-f35e-480f-b1aa-3320825d180b/volumes" Feb 15 21:28:17 crc kubenswrapper[4735]: I0215 21:28:17.131086 4735 scope.go:117] "RemoveContainer" containerID="5da3e571ad86b50bbb55d7d565f8277689d6fc25fa6163541ac1e110cd8263d3" Feb 15 21:28:17 crc kubenswrapper[4735]: I0215 21:28:17.131375 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/crc-debug-kpm99" Feb 15 21:28:51 crc kubenswrapper[4735]: I0215 21:28:51.040078 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d4fd7fbc4-nt694_c0df081d-fe76-415d-b26f-c67e2c97284f/barbican-api/0.log" Feb 15 21:28:51 crc kubenswrapper[4735]: I0215 21:28:51.246645 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d4fd7fbc4-nt694_c0df081d-fe76-415d-b26f-c67e2c97284f/barbican-api-log/0.log" Feb 15 21:28:51 crc kubenswrapper[4735]: I0215 21:28:51.306375 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f776895b6-v9kd8_8c4ffab0-9e1c-4ccb-841d-54c812217899/barbican-keystone-listener/0.log" Feb 15 21:28:51 crc kubenswrapper[4735]: I0215 21:28:51.369570 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f776895b6-v9kd8_8c4ffab0-9e1c-4ccb-841d-54c812217899/barbican-keystone-listener-log/0.log" Feb 15 21:28:51 crc kubenswrapper[4735]: I0215 21:28:51.569679 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b497d9d6c-4dqkk_29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2/barbican-worker-log/0.log" Feb 15 21:28:51 crc kubenswrapper[4735]: I0215 21:28:51.572755 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b497d9d6c-4dqkk_29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2/barbican-worker/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.089600 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr_b5df6771-b0ed-45d4-aeac-219bb79caf7a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.114733 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_09208797-cc26-4216-b27f-51b281eba74a/ceilometer-central-agent/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.172070 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_09208797-cc26-4216-b27f-51b281eba74a/ceilometer-notification-agent/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.278784 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_09208797-cc26-4216-b27f-51b281eba74a/proxy-httpd/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.347379 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_09208797-cc26-4216-b27f-51b281eba74a/sg-core/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.491827 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1f3720a1-390e-457d-a4c2-d758b04a90d3/cinder-api/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.546931 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1f3720a1-390e-457d-a4c2-d758b04a90d3/cinder-api-log/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.710560 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7464049d-a3f2-44c0-982d-222196019aa2/probe/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.747988 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7464049d-a3f2-44c0-982d-222196019aa2/cinder-scheduler/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.927905 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-bwldn_f62f1800-36ef-497f-94af-ca96c2a3f7e9/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:28:52 crc kubenswrapper[4735]: I0215 21:28:52.991617 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x_4a254833-483c-4f57-8407-2dad01349aec/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:28:53 crc kubenswrapper[4735]: I0215 21:28:53.154097 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-pnv2f_fd43082b-0db8-4620-9784-cc587548f737/init/0.log" Feb 15 21:28:53 crc kubenswrapper[4735]: I0215 21:28:53.746654 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-pnv2f_fd43082b-0db8-4620-9784-cc587548f737/init/0.log" Feb 15 21:28:54 crc kubenswrapper[4735]: I0215 21:28:54.053410 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg_a94a8020-82e8-4b37-add7-d9be8fc04c44/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:28:54 crc kubenswrapper[4735]: I0215 21:28:54.057644 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-pnv2f_fd43082b-0db8-4620-9784-cc587548f737/dnsmasq-dns/0.log" Feb 15 21:28:54 crc kubenswrapper[4735]: I0215 21:28:54.241184 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3a32ed00-6692-40df-af48-ecd34af2ffd2/glance-httpd/0.log" Feb 15 21:28:54 crc kubenswrapper[4735]: I0215 21:28:54.251184 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3a32ed00-6692-40df-af48-ecd34af2ffd2/glance-log/0.log" Feb 15 21:28:54 crc kubenswrapper[4735]: I0215 21:28:54.314259 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91998299-81cc-409a-a32d-2d2fa671d379/glance-httpd/0.log" Feb 15 21:28:54 crc kubenswrapper[4735]: I0215 21:28:54.477487 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91998299-81cc-409a-a32d-2d2fa671d379/glance-log/0.log" Feb 15 21:28:54 crc kubenswrapper[4735]: I0215 21:28:54.592575 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-c647fbf5b-7zs9r_970d5293-b621-45e1-a1e4-8cc176c9a148/horizon/2.log" Feb 15 21:28:54 crc kubenswrapper[4735]: I0215 21:28:54.685566 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-c647fbf5b-7zs9r_970d5293-b621-45e1-a1e4-8cc176c9a148/horizon/1.log" Feb 15 21:28:54 crc kubenswrapper[4735]: I0215 21:28:54.960438 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv_0ea9cfe8-7371-4d83-8e57-755d86809d46/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:28:55 crc kubenswrapper[4735]: I0215 21:28:55.085061 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-c647fbf5b-7zs9r_970d5293-b621-45e1-a1e4-8cc176c9a148/horizon-log/0.log" Feb 15 21:28:55 crc kubenswrapper[4735]: I0215 21:28:55.401586 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-c7dpc_38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:28:55 crc kubenswrapper[4735]: I0215 21:28:55.686594 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29519821-8vcgr_25286b3f-a5eb-49ce-a0dc-e2a8e99d9264/keystone-cron/0.log" Feb 15 21:28:55 crc kubenswrapper[4735]: I0215 21:28:55.762351 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7cb4db7f5b-mrtp9_33778124-7c75-4b6f-a2e1-c0ac0d8eda7e/keystone-api/0.log" Feb 15 21:28:55 crc kubenswrapper[4735]: I0215 21:28:55.784055 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04/kube-state-metrics/0.log" Feb 15 21:28:56 crc kubenswrapper[4735]: I0215 21:28:56.062644 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g_3e63486b-442e-477f-8692-cb3da5e648cd/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:28:56 crc kubenswrapper[4735]: I0215 21:28:56.434569 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs_82354908-424c-4069-abd7-f6b5ededdf13/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:28:56 crc kubenswrapper[4735]: I0215 21:28:56.600935 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-fdf549c7-xpft7_33cc02bd-6580-4d99-9c26-5f0c222e6461/neutron-httpd/0.log" Feb 15 21:28:56 crc kubenswrapper[4735]: I0215 21:28:56.897033 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-fdf549c7-xpft7_33cc02bd-6580-4d99-9c26-5f0c222e6461/neutron-api/0.log" Feb 15 21:28:57 crc kubenswrapper[4735]: I0215 21:28:57.354898 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d38fb2ae-a738-456f-b55e-65c70672dd2a/nova-cell0-conductor-conductor/0.log" Feb 15 21:28:57 crc kubenswrapper[4735]: I0215 21:28:57.548802 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_943d7959-9276-4940-8783-27492dfdf13e/nova-cell1-conductor-conductor/0.log" Feb 15 21:28:57 crc kubenswrapper[4735]: I0215 21:28:57.899652 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_cc100b91-f66d-43d2-8b63-d52a3c27769f/nova-cell1-novncproxy-novncproxy/0.log" Feb 15 21:28:58 crc kubenswrapper[4735]: I0215 21:28:57.999983 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a964a23e-2012-4d43-98c2-e0b15f9d9d74/nova-api-log/0.log" Feb 15 21:28:58 crc kubenswrapper[4735]: I0215 21:28:58.113350 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a964a23e-2012-4d43-98c2-e0b15f9d9d74/nova-api-api/0.log" Feb 15 21:28:58 crc kubenswrapper[4735]: I0215 21:28:58.281606 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-t2s42_01424332-5cb6-4959-90b3-eb98af1ee96e/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:28:58 crc kubenswrapper[4735]: I0215 21:28:58.325997 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_90c9c199-a27c-464a-8845-c2fd4e0b0fd9/nova-metadata-log/0.log" Feb 15 21:28:58 crc kubenswrapper[4735]: I0215 21:28:58.769480 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1f87d118-86d0-477e-a735-d2cd281b334a/mysql-bootstrap/0.log" Feb 15 21:28:58 crc kubenswrapper[4735]: I0215 21:28:58.992649 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d2066599-479d-43c3-86cb-528fc3922810/nova-scheduler-scheduler/0.log" Feb 15 21:28:59 crc kubenswrapper[4735]: I0215 21:28:59.055758 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1f87d118-86d0-477e-a735-d2cd281b334a/mysql-bootstrap/0.log" Feb 15 21:28:59 crc kubenswrapper[4735]: I0215 21:28:59.064067 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1f87d118-86d0-477e-a735-d2cd281b334a/galera/0.log" Feb 15 21:28:59 crc kubenswrapper[4735]: I0215 21:28:59.395724 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aa84f1ba-392a-4523-927c-405fab5bd619/mysql-bootstrap/0.log" Feb 15 21:28:59 crc kubenswrapper[4735]: I0215 21:28:59.631116 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aa84f1ba-392a-4523-927c-405fab5bd619/galera/0.log" Feb 15 21:28:59 crc kubenswrapper[4735]: I0215 21:28:59.674886 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aa84f1ba-392a-4523-927c-405fab5bd619/mysql-bootstrap/0.log" Feb 15 21:28:59 crc kubenswrapper[4735]: I0215 21:28:59.825318 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15/openstackclient/0.log" Feb 15 21:28:59 crc kubenswrapper[4735]: I0215 21:28:59.971467 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-8vw9j_3ff080f2-642e-4ff4-9554-5489daca64c4/openstack-network-exporter/0.log" Feb 15 21:29:00 crc kubenswrapper[4735]: I0215 21:29:00.011500 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_90c9c199-a27c-464a-8845-c2fd4e0b0fd9/nova-metadata-metadata/0.log" Feb 15 21:29:00 crc kubenswrapper[4735]: I0215 21:29:00.187257 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fgcfc_9ca538ce-5609-4a07-887c-b9757cd7ad37/ovsdb-server-init/0.log" Feb 15 21:29:00 crc kubenswrapper[4735]: I0215 21:29:00.405315 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fgcfc_9ca538ce-5609-4a07-887c-b9757cd7ad37/ovsdb-server-init/0.log" Feb 15 21:29:00 crc kubenswrapper[4735]: I0215 21:29:00.495295 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fgcfc_9ca538ce-5609-4a07-887c-b9757cd7ad37/ovsdb-server/0.log" Feb 15 21:29:00 crc kubenswrapper[4735]: I0215 21:29:00.539124 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fgcfc_9ca538ce-5609-4a07-887c-b9757cd7ad37/ovs-vswitchd/0.log" Feb 15 21:29:00 crc kubenswrapper[4735]: I0215 21:29:00.657379 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tk256_df76a192-fa00-4146-8f3a-8f7fa6798fec/ovn-controller/0.log" Feb 15 21:29:00 crc kubenswrapper[4735]: I0215 21:29:00.813990 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-qc8mw_5779edb8-01f9-4f91-81fe-9af70711a9bc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:29:00 crc kubenswrapper[4735]: I0215 21:29:00.990068 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d/openstack-network-exporter/0.log" Feb 15 21:29:01 crc kubenswrapper[4735]: I0215 21:29:01.057682 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d/ovn-northd/0.log" Feb 15 21:29:01 crc kubenswrapper[4735]: I0215 21:29:01.143605 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_77c1f16a-92e0-4c7a-a92c-12328f59f0b8/openstack-network-exporter/0.log" Feb 15 21:29:01 crc kubenswrapper[4735]: I0215 21:29:01.281176 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_77c1f16a-92e0-4c7a-a92c-12328f59f0b8/ovsdbserver-nb/0.log" Feb 15 21:29:01 crc kubenswrapper[4735]: I0215 21:29:01.394662 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8bfd475e-d45a-4d6e-8150-90da2bf08b6a/openstack-network-exporter/0.log" Feb 15 21:29:01 crc kubenswrapper[4735]: I0215 21:29:01.473785 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8bfd475e-d45a-4d6e-8150-90da2bf08b6a/ovsdbserver-sb/0.log" Feb 15 21:29:02 crc kubenswrapper[4735]: I0215 21:29:02.320873 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b1105ffb-a8a0-44fd-8679-171e016f43b1/setup-container/0.log" Feb 15 21:29:02 crc kubenswrapper[4735]: I0215 21:29:02.342339 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6c5c4f5b9b-sg99c_da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f/placement-api/0.log" Feb 15 21:29:02 crc kubenswrapper[4735]: I0215 21:29:02.349119 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6c5c4f5b9b-sg99c_da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f/placement-log/0.log" Feb 15 21:29:02 crc kubenswrapper[4735]: I0215 21:29:02.484994 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b1105ffb-a8a0-44fd-8679-171e016f43b1/setup-container/0.log" Feb 15 21:29:02 crc kubenswrapper[4735]: I0215 21:29:02.647392 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b1105ffb-a8a0-44fd-8679-171e016f43b1/rabbitmq/0.log" Feb 15 21:29:02 crc kubenswrapper[4735]: I0215 21:29:02.704836 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40/setup-container/0.log" Feb 15 21:29:02 crc kubenswrapper[4735]: I0215 21:29:02.795380 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40/setup-container/0.log" Feb 15 21:29:02 crc kubenswrapper[4735]: I0215 21:29:02.944065 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40/rabbitmq/0.log" Feb 15 21:29:02 crc kubenswrapper[4735]: I0215 21:29:02.977924 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf_7bbcab5d-3968-4e71-b53f-0c78e86b21cb/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:29:03 crc kubenswrapper[4735]: I0215 21:29:03.486927 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-df8c9_aad75417-7c89-443b-88ed-aad98d4109b2/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:29:03 crc kubenswrapper[4735]: I0215 21:29:03.541656 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2_e31b2a4c-8571-4a0e-868f-f86b757785ac/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:29:03 crc kubenswrapper[4735]: I0215 21:29:03.818066 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9wkf7_a7ab7547-fa81-411c-b9dd-fed0f819cd41/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:29:03 crc kubenswrapper[4735]: I0215 21:29:03.931516 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-frlr4_08f45e83-df1b-430c-a2af-92b9370912a5/ssh-known-hosts-edpm-deployment/0.log" Feb 15 21:29:04 crc kubenswrapper[4735]: I0215 21:29:04.095197 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b7b86987-zpgc8_843b8077-b23a-42f9-bf70-79767ce35b4f/proxy-server/0.log" Feb 15 21:29:04 crc kubenswrapper[4735]: I0215 21:29:04.185780 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-zrmz5_35e018a0-8bf3-4725-a229-5a4196d4ae97/swift-ring-rebalance/0.log" Feb 15 21:29:04 crc kubenswrapper[4735]: I0215 21:29:04.222248 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b7b86987-zpgc8_843b8077-b23a-42f9-bf70-79767ce35b4f/proxy-httpd/0.log" Feb 15 21:29:04 crc kubenswrapper[4735]: I0215 21:29:04.705579 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/account-reaper/0.log" Feb 15 21:29:04 crc kubenswrapper[4735]: I0215 21:29:04.706678 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/account-auditor/0.log" Feb 15 21:29:04 crc kubenswrapper[4735]: I0215 21:29:04.842002 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/account-server/0.log" Feb 15 21:29:04 crc kubenswrapper[4735]: I0215 21:29:04.941768 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/account-replicator/0.log" Feb 15 21:29:04 crc kubenswrapper[4735]: I0215 21:29:04.981176 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/container-auditor/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.065317 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/container-replicator/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.140745 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/container-server/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.303553 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-auditor/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.325907 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/container-updater/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.402423 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-expirer/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.428858 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-replicator/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.517536 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-server/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.528884 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-updater/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.594313 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/rsync/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.688580 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/swift-recon-cron/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.912783 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf_25a254b5-abc4-45fa-aba5-935d79ae8959/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:29:05 crc kubenswrapper[4735]: I0215 21:29:05.945858 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_abcf2b9d-8410-4a1b-be0e-3dde2a6e3383/tempest-tests-tempest-tests-runner/0.log" Feb 15 21:29:06 crc kubenswrapper[4735]: I0215 21:29:06.114249 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_97a6ddef-7ccf-4c1f-bc90-691f4511e4ad/test-operator-logs-container/0.log" Feb 15 21:29:06 crc kubenswrapper[4735]: I0215 21:29:06.191935 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch_6c7b61e8-5a6a-4651-bf3b-3f90ae858b40/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:29:18 crc kubenswrapper[4735]: I0215 21:29:18.828541 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6a34b649-4566-4a07-846e-e97886eec195/memcached/0.log" Feb 15 21:29:37 crc kubenswrapper[4735]: I0215 21:29:37.447212 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/util/0.log" Feb 15 21:29:37 crc kubenswrapper[4735]: I0215 21:29:37.658301 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/util/0.log" Feb 15 21:29:37 crc kubenswrapper[4735]: I0215 21:29:37.697099 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/pull/0.log" Feb 15 21:29:37 crc kubenswrapper[4735]: I0215 21:29:37.705758 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/pull/0.log" Feb 15 21:29:37 crc kubenswrapper[4735]: I0215 21:29:37.905998 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/util/0.log" Feb 15 21:29:37 crc kubenswrapper[4735]: I0215 21:29:37.981985 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/pull/0.log" Feb 15 21:29:37 crc kubenswrapper[4735]: I0215 21:29:37.982594 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/extract/0.log" Feb 15 21:29:38 crc kubenswrapper[4735]: I0215 21:29:38.888084 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-vddvx_99564ffe-1d8b-44b1-a625-4445d624a98f/manager/0.log" Feb 15 21:29:39 crc kubenswrapper[4735]: I0215 21:29:39.175690 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-gvnkh_92aeecb8-b9fa-4db8-916e-f2dd800d2da3/manager/0.log" Feb 15 21:29:39 crc kubenswrapper[4735]: I0215 21:29:39.368611 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-v4wpf_ab28d8d3-58f1-460c-a0ed-75f61ebc2c52/manager/0.log" Feb 15 21:29:39 crc kubenswrapper[4735]: I0215 21:29:39.641001 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-fslrr_31865132-118d-4f4c-bf0b-44ab13060882/manager/0.log" Feb 15 21:29:40 crc kubenswrapper[4735]: I0215 21:29:40.333737 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-rbhc6_2b6ae856-4502-44cd-ad23-21577792b6a0/manager/0.log" Feb 15 21:29:40 crc kubenswrapper[4735]: I0215 21:29:40.370722 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7676946c7c-5w649_79556e2d-fb3e-4a41-a85c-38005174ba50/manager/0.log" Feb 15 21:29:40 crc kubenswrapper[4735]: I0215 21:29:40.378790 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-v6nvk_98e57a69-0860-468e-898c-9ec42836d3d4/manager/0.log" Feb 15 21:29:40 crc kubenswrapper[4735]: I0215 21:29:40.800468 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-ghs7m_8802a6f8-206d-473d-854c-77e3c74389b5/manager/0.log" Feb 15 21:29:40 crc kubenswrapper[4735]: I0215 21:29:40.933755 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-mh89r_1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0/manager/0.log" Feb 15 21:29:41 crc kubenswrapper[4735]: I0215 21:29:41.238606 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-jwrfj_79e82794-8193-4eba-af98-f6e0ae43274b/manager/0.log" Feb 15 21:29:41 crc kubenswrapper[4735]: I0215 21:29:41.288537 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-4hlsd_1094ce7c-b104-40d4-8e0f-3ee695bfd6cd/manager/0.log" Feb 15 21:29:41 crc kubenswrapper[4735]: I0215 21:29:41.689723 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-t9nc4_d8c734e9-fd8c-459d-982f-53b0a706d0a0/manager/0.log" Feb 15 21:29:41 crc kubenswrapper[4735]: I0215 21:29:41.977724 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84966cf5c48cdg5_9549fdaf-fe4c-4da2-92db-db91beb95450/manager/0.log" Feb 15 21:29:42 crc kubenswrapper[4735]: I0215 21:29:42.377799 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-58498c75bf-2kzsp_e8c2edc3-d324-4b8e-9831-fcf924f6a655/operator/0.log" Feb 15 21:29:42 crc kubenswrapper[4735]: I0215 21:29:42.587746 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-ldbt6_32c1d680-92cf-4c49-93b9-17bb17aaf623/registry-server/0.log" Feb 15 21:29:42 crc kubenswrapper[4735]: I0215 21:29:42.949354 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-f8s72_1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea/manager/0.log" Feb 15 21:29:43 crc kubenswrapper[4735]: I0215 21:29:43.208936 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-9trcb_d8bf1d9d-df22-4b17-91b7-833052b5a337/manager/0.log" Feb 15 21:29:43 crc kubenswrapper[4735]: I0215 21:29:43.529043 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-9z8vz_e57b4eea-5c8f-4caa-9644-0994108974da/operator/0.log" Feb 15 21:29:43 crc kubenswrapper[4735]: I0215 21:29:43.776365 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-vbr6n_14de9f25-fdbd-4d44-9f2c-440b256c4faf/manager/0.log" Feb 15 21:29:44 crc kubenswrapper[4735]: I0215 21:29:44.106546 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-lfj2m_4fb2a283-35ca-4391-8e59-79a8fa7bb4f0/manager/0.log" Feb 15 21:29:44 crc kubenswrapper[4735]: I0215 21:29:44.263363 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-r6bz7_3fa6a777-e009-4e48-b05d-ac9b34acd21c/manager/0.log" Feb 15 21:29:44 crc kubenswrapper[4735]: I0215 21:29:44.298760 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79f9b74c8-4b426_6fad7194-6e10-4dc9-ac3c-241722357e55/manager/0.log" Feb 15 21:29:44 crc kubenswrapper[4735]: I0215 21:29:44.319984 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-w8vz9_c7d8cf5f-13c8-4857-96d2-0da2c88ed221/manager/0.log" Feb 15 21:29:44 crc kubenswrapper[4735]: I0215 21:29:44.528493 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-d65l6_26e7f3b0-6f34-491e-917a-e545d320ee89/manager/0.log" Feb 15 21:29:48 crc kubenswrapper[4735]: I0215 21:29:48.222779 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-9hc4f_b9be4a5f-fbec-4e52-bb00-cbc08576e1fd/manager/0.log" Feb 15 21:29:49 crc kubenswrapper[4735]: I0215 21:29:49.679760 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:29:49 crc kubenswrapper[4735]: I0215 21:29:49.680087 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.182569 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc"] Feb 15 21:30:00 crc kubenswrapper[4735]: E0215 21:30:00.183447 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc62a3ab-f35e-480f-b1aa-3320825d180b" containerName="container-00" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.183458 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc62a3ab-f35e-480f-b1aa-3320825d180b" containerName="container-00" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.183642 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc62a3ab-f35e-480f-b1aa-3320825d180b" containerName="container-00" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.185524 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.189992 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.195327 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc"] Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.196350 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.218773 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rnxp\" (UniqueName: \"kubernetes.io/projected/a4e5bca7-43e4-4efd-899f-a73040d0bda8-kube-api-access-8rnxp\") pod \"collect-profiles-29519850-tc6rc\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.218984 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4e5bca7-43e4-4efd-899f-a73040d0bda8-secret-volume\") pod \"collect-profiles-29519850-tc6rc\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.219014 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4e5bca7-43e4-4efd-899f-a73040d0bda8-config-volume\") pod \"collect-profiles-29519850-tc6rc\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.319927 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4e5bca7-43e4-4efd-899f-a73040d0bda8-secret-volume\") pod \"collect-profiles-29519850-tc6rc\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.320009 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4e5bca7-43e4-4efd-899f-a73040d0bda8-config-volume\") pod \"collect-profiles-29519850-tc6rc\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.320060 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rnxp\" (UniqueName: \"kubernetes.io/projected/a4e5bca7-43e4-4efd-899f-a73040d0bda8-kube-api-access-8rnxp\") pod \"collect-profiles-29519850-tc6rc\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.321771 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4e5bca7-43e4-4efd-899f-a73040d0bda8-config-volume\") pod \"collect-profiles-29519850-tc6rc\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.332282 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4e5bca7-43e4-4efd-899f-a73040d0bda8-secret-volume\") pod \"collect-profiles-29519850-tc6rc\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.336456 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rnxp\" (UniqueName: \"kubernetes.io/projected/a4e5bca7-43e4-4efd-899f-a73040d0bda8-kube-api-access-8rnxp\") pod \"collect-profiles-29519850-tc6rc\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.536783 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:00 crc kubenswrapper[4735]: I0215 21:30:00.964432 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc"] Feb 15 21:30:01 crc kubenswrapper[4735]: I0215 21:30:01.053155 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" event={"ID":"a4e5bca7-43e4-4efd-899f-a73040d0bda8","Type":"ContainerStarted","Data":"e37eb06833593d73509bcb4091ee37e36b75bbd8c3b0493c3f8a370e443bc2de"} Feb 15 21:30:02 crc kubenswrapper[4735]: I0215 21:30:02.063560 4735 generic.go:334] "Generic (PLEG): container finished" podID="a4e5bca7-43e4-4efd-899f-a73040d0bda8" containerID="5cc11d8697e79afc7b65569e3711da74c23e4dacc1659611e3e0b521efad839b" exitCode=0 Feb 15 21:30:02 crc kubenswrapper[4735]: I0215 21:30:02.063878 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" event={"ID":"a4e5bca7-43e4-4efd-899f-a73040d0bda8","Type":"ContainerDied","Data":"5cc11d8697e79afc7b65569e3711da74c23e4dacc1659611e3e0b521efad839b"} Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.712583 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.886524 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rnxp\" (UniqueName: \"kubernetes.io/projected/a4e5bca7-43e4-4efd-899f-a73040d0bda8-kube-api-access-8rnxp\") pod \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.886576 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4e5bca7-43e4-4efd-899f-a73040d0bda8-secret-volume\") pod \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.886612 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4e5bca7-43e4-4efd-899f-a73040d0bda8-config-volume\") pod \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\" (UID: \"a4e5bca7-43e4-4efd-899f-a73040d0bda8\") " Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.887693 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4e5bca7-43e4-4efd-899f-a73040d0bda8-config-volume" (OuterVolumeSpecName: "config-volume") pod "a4e5bca7-43e4-4efd-899f-a73040d0bda8" (UID: "a4e5bca7-43e4-4efd-899f-a73040d0bda8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.894195 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4e5bca7-43e4-4efd-899f-a73040d0bda8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a4e5bca7-43e4-4efd-899f-a73040d0bda8" (UID: "a4e5bca7-43e4-4efd-899f-a73040d0bda8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.895189 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e5bca7-43e4-4efd-899f-a73040d0bda8-kube-api-access-8rnxp" (OuterVolumeSpecName: "kube-api-access-8rnxp") pod "a4e5bca7-43e4-4efd-899f-a73040d0bda8" (UID: "a4e5bca7-43e4-4efd-899f-a73040d0bda8"). InnerVolumeSpecName "kube-api-access-8rnxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.989643 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rnxp\" (UniqueName: \"kubernetes.io/projected/a4e5bca7-43e4-4efd-899f-a73040d0bda8-kube-api-access-8rnxp\") on node \"crc\" DevicePath \"\"" Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.989686 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4e5bca7-43e4-4efd-899f-a73040d0bda8-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 15 21:30:03 crc kubenswrapper[4735]: I0215 21:30:03.989698 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4e5bca7-43e4-4efd-899f-a73040d0bda8-config-volume\") on node \"crc\" DevicePath \"\"" Feb 15 21:30:04 crc kubenswrapper[4735]: I0215 21:30:04.080388 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" event={"ID":"a4e5bca7-43e4-4efd-899f-a73040d0bda8","Type":"ContainerDied","Data":"e37eb06833593d73509bcb4091ee37e36b75bbd8c3b0493c3f8a370e443bc2de"} Feb 15 21:30:04 crc kubenswrapper[4735]: I0215 21:30:04.080432 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e37eb06833593d73509bcb4091ee37e36b75bbd8c3b0493c3f8a370e443bc2de" Feb 15 21:30:04 crc kubenswrapper[4735]: I0215 21:30:04.080481 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519850-tc6rc" Feb 15 21:30:04 crc kubenswrapper[4735]: I0215 21:30:04.800500 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586"] Feb 15 21:30:04 crc kubenswrapper[4735]: I0215 21:30:04.812753 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519805-6b586"] Feb 15 21:30:04 crc kubenswrapper[4735]: I0215 21:30:04.900286 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0b06b04-ee13-4172-99b2-6c5552365399" path="/var/lib/kubelet/pods/e0b06b04-ee13-4172-99b2-6c5552365399/volumes" Feb 15 21:30:09 crc kubenswrapper[4735]: I0215 21:30:09.087815 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-4wxcd_3036a2ba-e736-416a-bcf1-684d246cd3f1/control-plane-machine-set-operator/0.log" Feb 15 21:30:09 crc kubenswrapper[4735]: I0215 21:30:09.286695 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rjhgs_34482250-9472-45fe-aab5-485c7f129bca/kube-rbac-proxy/0.log" Feb 15 21:30:09 crc kubenswrapper[4735]: I0215 21:30:09.330109 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rjhgs_34482250-9472-45fe-aab5-485c7f129bca/machine-api-operator/0.log" Feb 15 21:30:19 crc kubenswrapper[4735]: I0215 21:30:19.680196 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:30:19 crc kubenswrapper[4735]: I0215 21:30:19.680714 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:30:24 crc kubenswrapper[4735]: I0215 21:30:24.006966 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-2jh8n_56f25e9b-225b-433d-af08-bdd6bb37e6e4/cert-manager-controller/0.log" Feb 15 21:30:24 crc kubenswrapper[4735]: I0215 21:30:24.266331 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-88v64_e04e2b12-a2ee-4ba4-acb4-0276c777a066/cert-manager-cainjector/0.log" Feb 15 21:30:24 crc kubenswrapper[4735]: I0215 21:30:24.300920 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-jvntw_468178fa-6b3c-41b4-9bf2-45178ebfb3c2/cert-manager-webhook/0.log" Feb 15 21:30:38 crc kubenswrapper[4735]: I0215 21:30:38.713742 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-6q5t8_4d5d6279-ce2c-4382-900c-77563978d63a/nmstate-console-plugin/0.log" Feb 15 21:30:38 crc kubenswrapper[4735]: I0215 21:30:38.934017 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5grq7_ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e/nmstate-handler/0.log" Feb 15 21:30:38 crc kubenswrapper[4735]: I0215 21:30:38.942001 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-4gnfl_6291dd59-7c75-4426-86d1-063a72855309/kube-rbac-proxy/0.log" Feb 15 21:30:39 crc kubenswrapper[4735]: I0215 21:30:39.072206 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-4gnfl_6291dd59-7c75-4426-86d1-063a72855309/nmstate-metrics/0.log" Feb 15 21:30:39 crc kubenswrapper[4735]: I0215 21:30:39.211030 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-vc7vf_16c72a4d-04eb-4a5f-af25-439d36f5c9de/nmstate-operator/0.log" Feb 15 21:30:39 crc kubenswrapper[4735]: I0215 21:30:39.327321 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-7gkdk_cecb1f2e-e07b-4ea1-924d-7802fa40fd6b/nmstate-webhook/0.log" Feb 15 21:30:49 crc kubenswrapper[4735]: I0215 21:30:49.680374 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:30:49 crc kubenswrapper[4735]: I0215 21:30:49.681088 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:30:49 crc kubenswrapper[4735]: I0215 21:30:49.681163 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 21:30:49 crc kubenswrapper[4735]: I0215 21:30:49.682058 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 21:30:49 crc kubenswrapper[4735]: I0215 21:30:49.682126 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" gracePeriod=600 Feb 15 21:30:49 crc kubenswrapper[4735]: E0215 21:30:49.799493 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:30:50 crc kubenswrapper[4735]: I0215 21:30:50.456966 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" exitCode=0 Feb 15 21:30:50 crc kubenswrapper[4735]: I0215 21:30:50.457010 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91"} Feb 15 21:30:50 crc kubenswrapper[4735]: I0215 21:30:50.457044 4735 scope.go:117] "RemoveContainer" containerID="5402e50ecca53aef653cfbc5bd98553d9acbbcbcfa2fcbeb5b89235c75b76baf" Feb 15 21:30:50 crc kubenswrapper[4735]: I0215 21:30:50.457728 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:30:50 crc kubenswrapper[4735]: E0215 21:30:50.458178 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:30:53 crc kubenswrapper[4735]: I0215 21:30:53.472554 4735 scope.go:117] "RemoveContainer" containerID="f3f1ee0cfa7e32ca2f5f444fc01874fd51e9fe0b6403263e19089cc8771530d4" Feb 15 21:31:05 crc kubenswrapper[4735]: I0215 21:31:05.886840 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:31:05 crc kubenswrapper[4735]: E0215 21:31:05.887537 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:31:12 crc kubenswrapper[4735]: I0215 21:31:12.571442 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-7xh6x_ca5b18f3-83a4-497d-913b-e0539a6991b5/kube-rbac-proxy/0.log" Feb 15 21:31:12 crc kubenswrapper[4735]: I0215 21:31:12.634963 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-7xh6x_ca5b18f3-83a4-497d-913b-e0539a6991b5/controller/0.log" Feb 15 21:31:12 crc kubenswrapper[4735]: I0215 21:31:12.867777 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-frr-files/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.089397 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-metrics/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.103755 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-reloader/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.144449 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-reloader/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.156241 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-frr-files/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.326153 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-metrics/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.327377 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-reloader/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.379972 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-metrics/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.433161 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-frr-files/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.663401 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/controller/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.666337 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-frr-files/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.697423 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-metrics/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.709462 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-reloader/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.914940 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/frr-metrics/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.926072 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/kube-rbac-proxy/0.log" Feb 15 21:31:13 crc kubenswrapper[4735]: I0215 21:31:13.960351 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/kube-rbac-proxy-frr/0.log" Feb 15 21:31:14 crc kubenswrapper[4735]: I0215 21:31:14.198834 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/reloader/0.log" Feb 15 21:31:14 crc kubenswrapper[4735]: I0215 21:31:14.313872 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-7km29_99caf845-7f1f-4a29-a8e5-c939ffb15c9e/frr-k8s-webhook-server/0.log" Feb 15 21:31:14 crc kubenswrapper[4735]: I0215 21:31:14.637643 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79df65bf49-hzgd7_c27da055-adee-4524-93ce-e2e83cf5966a/manager/0.log" Feb 15 21:31:14 crc kubenswrapper[4735]: I0215 21:31:14.792350 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7f47cf9495-7nwhm_3516fdcd-aa1a-48ea-9f44-b928ca524e28/webhook-server/0.log" Feb 15 21:31:15 crc kubenswrapper[4735]: I0215 21:31:15.035635 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/frr/0.log" Feb 15 21:31:15 crc kubenswrapper[4735]: I0215 21:31:15.047650 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9tpjm_ec4ea5be-56b7-42e2-b462-a880d12145af/kube-rbac-proxy/0.log" Feb 15 21:31:15 crc kubenswrapper[4735]: I0215 21:31:15.802305 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9tpjm_ec4ea5be-56b7-42e2-b462-a880d12145af/speaker/0.log" Feb 15 21:31:18 crc kubenswrapper[4735]: I0215 21:31:18.936477 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:31:18 crc kubenswrapper[4735]: E0215 21:31:18.936942 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:31:32 crc kubenswrapper[4735]: I0215 21:31:32.191735 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/util/0.log" Feb 15 21:31:32 crc kubenswrapper[4735]: I0215 21:31:32.393737 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/util/0.log" Feb 15 21:31:32 crc kubenswrapper[4735]: I0215 21:31:32.407811 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/pull/0.log" Feb 15 21:31:32 crc kubenswrapper[4735]: I0215 21:31:32.417615 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/pull/0.log" Feb 15 21:31:32 crc kubenswrapper[4735]: I0215 21:31:32.582241 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/pull/0.log" Feb 15 21:31:32 crc kubenswrapper[4735]: I0215 21:31:32.619682 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/extract/0.log" Feb 15 21:31:32 crc kubenswrapper[4735]: I0215 21:31:32.626889 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/util/0.log" Feb 15 21:31:32 crc kubenswrapper[4735]: I0215 21:31:32.796664 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-utilities/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.050534 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-utilities/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.055466 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-content/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.092469 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-content/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.291668 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-content/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.295843 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-utilities/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.608468 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-utilities/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.709303 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/registry-server/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.840465 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-content/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.846559 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-utilities/0.log" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.887740 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:31:33 crc kubenswrapper[4735]: E0215 21:31:33.887929 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:31:33 crc kubenswrapper[4735]: I0215 21:31:33.919720 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-content/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.114635 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-utilities/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.128435 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-content/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.380046 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/util/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.626732 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/registry-server/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.695526 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/util/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.698208 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/pull/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.736670 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/pull/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.895801 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/pull/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.940612 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/util/0.log" Feb 15 21:31:34 crc kubenswrapper[4735]: I0215 21:31:34.951341 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/extract/0.log" Feb 15 21:31:35 crc kubenswrapper[4735]: I0215 21:31:35.108723 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ff4s6_02a0669b-e47e-4f62-851c-a776d1f43b12/marketplace-operator/0.log" Feb 15 21:31:35 crc kubenswrapper[4735]: I0215 21:31:35.221451 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-utilities/0.log" Feb 15 21:31:35 crc kubenswrapper[4735]: I0215 21:31:35.403908 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-content/0.log" Feb 15 21:31:35 crc kubenswrapper[4735]: I0215 21:31:35.441037 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-utilities/0.log" Feb 15 21:31:35 crc kubenswrapper[4735]: I0215 21:31:35.483089 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-content/0.log" Feb 15 21:31:36 crc kubenswrapper[4735]: I0215 21:31:36.043167 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-content/0.log" Feb 15 21:31:36 crc kubenswrapper[4735]: I0215 21:31:36.044884 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-utilities/0.log" Feb 15 21:31:36 crc kubenswrapper[4735]: I0215 21:31:36.167500 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/registry-server/0.log" Feb 15 21:31:36 crc kubenswrapper[4735]: I0215 21:31:36.358577 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-utilities/0.log" Feb 15 21:31:36 crc kubenswrapper[4735]: I0215 21:31:36.556147 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-content/0.log" Feb 15 21:31:36 crc kubenswrapper[4735]: I0215 21:31:36.605796 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-content/0.log" Feb 15 21:31:36 crc kubenswrapper[4735]: I0215 21:31:36.622453 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-utilities/0.log" Feb 15 21:31:36 crc kubenswrapper[4735]: I0215 21:31:36.823689 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-content/0.log" Feb 15 21:31:36 crc kubenswrapper[4735]: I0215 21:31:36.827141 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-utilities/0.log" Feb 15 21:31:37 crc kubenswrapper[4735]: I0215 21:31:37.398671 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/registry-server/0.log" Feb 15 21:31:48 crc kubenswrapper[4735]: I0215 21:31:48.887368 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:31:48 crc kubenswrapper[4735]: E0215 21:31:48.888036 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:31:55 crc kubenswrapper[4735]: E0215 21:31:55.693818 4735 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.82:53466->38.102.83.82:38695: read tcp 38.102.83.82:53466->38.102.83.82:38695: read: connection reset by peer Feb 15 21:31:59 crc kubenswrapper[4735]: I0215 21:31:59.887356 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:31:59 crc kubenswrapper[4735]: E0215 21:31:59.888090 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:32:11 crc kubenswrapper[4735]: I0215 21:32:11.886823 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:32:11 crc kubenswrapper[4735]: E0215 21:32:11.888344 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.314555 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7q9rt"] Feb 15 21:32:15 crc kubenswrapper[4735]: E0215 21:32:15.317580 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e5bca7-43e4-4efd-899f-a73040d0bda8" containerName="collect-profiles" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.317621 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e5bca7-43e4-4efd-899f-a73040d0bda8" containerName="collect-profiles" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.317828 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e5bca7-43e4-4efd-899f-a73040d0bda8" containerName="collect-profiles" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.319630 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.331365 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7q9rt"] Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.423594 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-utilities\") pod \"certified-operators-7q9rt\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.423892 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlncs\" (UniqueName: \"kubernetes.io/projected/1c89388f-ca41-4533-9e8e-540651ad7a0a-kube-api-access-vlncs\") pod \"certified-operators-7q9rt\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.424058 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-catalog-content\") pod \"certified-operators-7q9rt\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.525623 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-catalog-content\") pod \"certified-operators-7q9rt\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.525746 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-utilities\") pod \"certified-operators-7q9rt\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.525804 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlncs\" (UniqueName: \"kubernetes.io/projected/1c89388f-ca41-4533-9e8e-540651ad7a0a-kube-api-access-vlncs\") pod \"certified-operators-7q9rt\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.526590 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-utilities\") pod \"certified-operators-7q9rt\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.526658 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-catalog-content\") pod \"certified-operators-7q9rt\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.545619 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlncs\" (UniqueName: \"kubernetes.io/projected/1c89388f-ca41-4533-9e8e-540651ad7a0a-kube-api-access-vlncs\") pod \"certified-operators-7q9rt\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:15 crc kubenswrapper[4735]: I0215 21:32:15.638820 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:16 crc kubenswrapper[4735]: I0215 21:32:16.321010 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7q9rt"] Feb 15 21:32:17 crc kubenswrapper[4735]: I0215 21:32:17.181839 4735 generic.go:334] "Generic (PLEG): container finished" podID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerID="7a640a5d470807dffee8ac58f9b72e4d740538894a517bc59fd595021f9514b1" exitCode=0 Feb 15 21:32:17 crc kubenswrapper[4735]: I0215 21:32:17.181929 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7q9rt" event={"ID":"1c89388f-ca41-4533-9e8e-540651ad7a0a","Type":"ContainerDied","Data":"7a640a5d470807dffee8ac58f9b72e4d740538894a517bc59fd595021f9514b1"} Feb 15 21:32:17 crc kubenswrapper[4735]: I0215 21:32:17.182159 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7q9rt" event={"ID":"1c89388f-ca41-4533-9e8e-540651ad7a0a","Type":"ContainerStarted","Data":"9d4bb26d666b8d0ca05d7b1420fb480f5b4fd4e5e04cf810cbfb009428f57bd6"} Feb 15 21:32:17 crc kubenswrapper[4735]: I0215 21:32:17.194845 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 21:32:19 crc kubenswrapper[4735]: I0215 21:32:19.199472 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7q9rt" event={"ID":"1c89388f-ca41-4533-9e8e-540651ad7a0a","Type":"ContainerStarted","Data":"d05d52f053a49e5b54af0bffbc7103981c09813b966a3ea894b74c797d9e5114"} Feb 15 21:32:20 crc kubenswrapper[4735]: I0215 21:32:20.209802 4735 generic.go:334] "Generic (PLEG): container finished" podID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerID="d05d52f053a49e5b54af0bffbc7103981c09813b966a3ea894b74c797d9e5114" exitCode=0 Feb 15 21:32:20 crc kubenswrapper[4735]: I0215 21:32:20.209861 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7q9rt" event={"ID":"1c89388f-ca41-4533-9e8e-540651ad7a0a","Type":"ContainerDied","Data":"d05d52f053a49e5b54af0bffbc7103981c09813b966a3ea894b74c797d9e5114"} Feb 15 21:32:21 crc kubenswrapper[4735]: I0215 21:32:21.222329 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7q9rt" event={"ID":"1c89388f-ca41-4533-9e8e-540651ad7a0a","Type":"ContainerStarted","Data":"f7809f4bfde5641d7f8c4e9593accacd595df8148e335a49ae09462ddb0918e8"} Feb 15 21:32:23 crc kubenswrapper[4735]: I0215 21:32:23.886633 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:32:23 crc kubenswrapper[4735]: E0215 21:32:23.887473 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:32:25 crc kubenswrapper[4735]: I0215 21:32:25.639968 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:25 crc kubenswrapper[4735]: I0215 21:32:25.640331 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:25 crc kubenswrapper[4735]: I0215 21:32:25.696901 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:25 crc kubenswrapper[4735]: I0215 21:32:25.725321 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7q9rt" podStartSLOduration=7.284123786 podStartE2EDuration="10.725301166s" podCreationTimestamp="2026-02-15 21:32:15 +0000 UTC" firstStartedPulling="2026-02-15 21:32:17.194561241 +0000 UTC m=+4545.060576864" lastFinishedPulling="2026-02-15 21:32:20.635738621 +0000 UTC m=+4548.501754244" observedRunningTime="2026-02-15 21:32:21.252141834 +0000 UTC m=+4549.118157457" watchObservedRunningTime="2026-02-15 21:32:25.725301166 +0000 UTC m=+4553.591316789" Feb 15 21:32:27 crc kubenswrapper[4735]: I0215 21:32:27.006444 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:27 crc kubenswrapper[4735]: I0215 21:32:27.056785 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7q9rt"] Feb 15 21:32:28 crc kubenswrapper[4735]: I0215 21:32:28.281004 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7q9rt" podUID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerName="registry-server" containerID="cri-o://f7809f4bfde5641d7f8c4e9593accacd595df8148e335a49ae09462ddb0918e8" gracePeriod=2 Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.303285 4735 generic.go:334] "Generic (PLEG): container finished" podID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerID="f7809f4bfde5641d7f8c4e9593accacd595df8148e335a49ae09462ddb0918e8" exitCode=0 Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.303841 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7q9rt" event={"ID":"1c89388f-ca41-4533-9e8e-540651ad7a0a","Type":"ContainerDied","Data":"f7809f4bfde5641d7f8c4e9593accacd595df8148e335a49ae09462ddb0918e8"} Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.492005 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.590925 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-utilities\") pod \"1c89388f-ca41-4533-9e8e-540651ad7a0a\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.591187 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-catalog-content\") pod \"1c89388f-ca41-4533-9e8e-540651ad7a0a\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.591210 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlncs\" (UniqueName: \"kubernetes.io/projected/1c89388f-ca41-4533-9e8e-540651ad7a0a-kube-api-access-vlncs\") pod \"1c89388f-ca41-4533-9e8e-540651ad7a0a\" (UID: \"1c89388f-ca41-4533-9e8e-540651ad7a0a\") " Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.593662 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-utilities" (OuterVolumeSpecName: "utilities") pod "1c89388f-ca41-4533-9e8e-540651ad7a0a" (UID: "1c89388f-ca41-4533-9e8e-540651ad7a0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.600193 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c89388f-ca41-4533-9e8e-540651ad7a0a-kube-api-access-vlncs" (OuterVolumeSpecName: "kube-api-access-vlncs") pod "1c89388f-ca41-4533-9e8e-540651ad7a0a" (UID: "1c89388f-ca41-4533-9e8e-540651ad7a0a"). InnerVolumeSpecName "kube-api-access-vlncs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.649862 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c89388f-ca41-4533-9e8e-540651ad7a0a" (UID: "1c89388f-ca41-4533-9e8e-540651ad7a0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.693633 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.693670 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlncs\" (UniqueName: \"kubernetes.io/projected/1c89388f-ca41-4533-9e8e-540651ad7a0a-kube-api-access-vlncs\") on node \"crc\" DevicePath \"\"" Feb 15 21:32:29 crc kubenswrapper[4735]: I0215 21:32:29.693683 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c89388f-ca41-4533-9e8e-540651ad7a0a-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:32:30 crc kubenswrapper[4735]: I0215 21:32:30.317277 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7q9rt" event={"ID":"1c89388f-ca41-4533-9e8e-540651ad7a0a","Type":"ContainerDied","Data":"9d4bb26d666b8d0ca05d7b1420fb480f5b4fd4e5e04cf810cbfb009428f57bd6"} Feb 15 21:32:30 crc kubenswrapper[4735]: I0215 21:32:30.317324 4735 scope.go:117] "RemoveContainer" containerID="f7809f4bfde5641d7f8c4e9593accacd595df8148e335a49ae09462ddb0918e8" Feb 15 21:32:30 crc kubenswrapper[4735]: I0215 21:32:30.317359 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7q9rt" Feb 15 21:32:30 crc kubenswrapper[4735]: I0215 21:32:30.343854 4735 scope.go:117] "RemoveContainer" containerID="d05d52f053a49e5b54af0bffbc7103981c09813b966a3ea894b74c797d9e5114" Feb 15 21:32:30 crc kubenswrapper[4735]: I0215 21:32:30.373410 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7q9rt"] Feb 15 21:32:30 crc kubenswrapper[4735]: I0215 21:32:30.379767 4735 scope.go:117] "RemoveContainer" containerID="7a640a5d470807dffee8ac58f9b72e4d740538894a517bc59fd595021f9514b1" Feb 15 21:32:30 crc kubenswrapper[4735]: I0215 21:32:30.388877 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7q9rt"] Feb 15 21:32:30 crc kubenswrapper[4735]: I0215 21:32:30.902364 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c89388f-ca41-4533-9e8e-540651ad7a0a" path="/var/lib/kubelet/pods/1c89388f-ca41-4533-9e8e-540651ad7a0a/volumes" Feb 15 21:32:35 crc kubenswrapper[4735]: I0215 21:32:35.886739 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:32:35 crc kubenswrapper[4735]: E0215 21:32:35.887561 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:32:50 crc kubenswrapper[4735]: I0215 21:32:50.886865 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:32:50 crc kubenswrapper[4735]: E0215 21:32:50.887626 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:33:03 crc kubenswrapper[4735]: I0215 21:33:03.887504 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:33:03 crc kubenswrapper[4735]: E0215 21:33:03.888837 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.480986 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cdqzr"] Feb 15 21:33:06 crc kubenswrapper[4735]: E0215 21:33:06.481766 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerName="registry-server" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.481783 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerName="registry-server" Feb 15 21:33:06 crc kubenswrapper[4735]: E0215 21:33:06.482113 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerName="extract-utilities" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.482129 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerName="extract-utilities" Feb 15 21:33:06 crc kubenswrapper[4735]: E0215 21:33:06.482163 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerName="extract-content" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.482177 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerName="extract-content" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.482481 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c89388f-ca41-4533-9e8e-540651ad7a0a" containerName="registry-server" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.484500 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.491287 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cdqzr"] Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.604829 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-utilities\") pod \"community-operators-cdqzr\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.604869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-catalog-content\") pod \"community-operators-cdqzr\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.604922 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxfk4\" (UniqueName: \"kubernetes.io/projected/63e0cc56-a704-4b36-a0dd-628b41f05569-kube-api-access-hxfk4\") pod \"community-operators-cdqzr\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.706767 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-utilities\") pod \"community-operators-cdqzr\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.706820 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-catalog-content\") pod \"community-operators-cdqzr\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.706897 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxfk4\" (UniqueName: \"kubernetes.io/projected/63e0cc56-a704-4b36-a0dd-628b41f05569-kube-api-access-hxfk4\") pod \"community-operators-cdqzr\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.707748 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-utilities\") pod \"community-operators-cdqzr\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.708038 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-catalog-content\") pod \"community-operators-cdqzr\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.736284 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxfk4\" (UniqueName: \"kubernetes.io/projected/63e0cc56-a704-4b36-a0dd-628b41f05569-kube-api-access-hxfk4\") pod \"community-operators-cdqzr\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:06 crc kubenswrapper[4735]: I0215 21:33:06.811759 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:07 crc kubenswrapper[4735]: I0215 21:33:07.901550 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cdqzr"] Feb 15 21:33:08 crc kubenswrapper[4735]: W0215 21:33:08.495453 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63e0cc56_a704_4b36_a0dd_628b41f05569.slice/crio-b05234da8bce7576b7ccc91a426b8339700e46959d439985f786ae7616217c21 WatchSource:0}: Error finding container b05234da8bce7576b7ccc91a426b8339700e46959d439985f786ae7616217c21: Status 404 returned error can't find the container with id b05234da8bce7576b7ccc91a426b8339700e46959d439985f786ae7616217c21 Feb 15 21:33:08 crc kubenswrapper[4735]: I0215 21:33:08.725397 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqzr" event={"ID":"63e0cc56-a704-4b36-a0dd-628b41f05569","Type":"ContainerStarted","Data":"fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be"} Feb 15 21:33:08 crc kubenswrapper[4735]: I0215 21:33:08.725769 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqzr" event={"ID":"63e0cc56-a704-4b36-a0dd-628b41f05569","Type":"ContainerStarted","Data":"b05234da8bce7576b7ccc91a426b8339700e46959d439985f786ae7616217c21"} Feb 15 21:33:09 crc kubenswrapper[4735]: I0215 21:33:09.736547 4735 generic.go:334] "Generic (PLEG): container finished" podID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerID="fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be" exitCode=0 Feb 15 21:33:09 crc kubenswrapper[4735]: I0215 21:33:09.736596 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqzr" event={"ID":"63e0cc56-a704-4b36-a0dd-628b41f05569","Type":"ContainerDied","Data":"fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be"} Feb 15 21:33:10 crc kubenswrapper[4735]: I0215 21:33:10.755038 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqzr" event={"ID":"63e0cc56-a704-4b36-a0dd-628b41f05569","Type":"ContainerStarted","Data":"2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b"} Feb 15 21:33:11 crc kubenswrapper[4735]: I0215 21:33:11.768088 4735 generic.go:334] "Generic (PLEG): container finished" podID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerID="2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b" exitCode=0 Feb 15 21:33:11 crc kubenswrapper[4735]: I0215 21:33:11.768691 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqzr" event={"ID":"63e0cc56-a704-4b36-a0dd-628b41f05569","Type":"ContainerDied","Data":"2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b"} Feb 15 21:33:12 crc kubenswrapper[4735]: I0215 21:33:12.780076 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqzr" event={"ID":"63e0cc56-a704-4b36-a0dd-628b41f05569","Type":"ContainerStarted","Data":"22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df"} Feb 15 21:33:12 crc kubenswrapper[4735]: I0215 21:33:12.824552 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cdqzr" podStartSLOduration=4.420656369 podStartE2EDuration="6.824526962s" podCreationTimestamp="2026-02-15 21:33:06 +0000 UTC" firstStartedPulling="2026-02-15 21:33:09.73885113 +0000 UTC m=+4597.604866753" lastFinishedPulling="2026-02-15 21:33:12.142721713 +0000 UTC m=+4600.008737346" observedRunningTime="2026-02-15 21:33:12.802380918 +0000 UTC m=+4600.668396551" watchObservedRunningTime="2026-02-15 21:33:12.824526962 +0000 UTC m=+4600.690542595" Feb 15 21:33:15 crc kubenswrapper[4735]: I0215 21:33:15.888351 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:33:15 crc kubenswrapper[4735]: E0215 21:33:15.888917 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:33:16 crc kubenswrapper[4735]: I0215 21:33:16.811942 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:16 crc kubenswrapper[4735]: I0215 21:33:16.812257 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:16 crc kubenswrapper[4735]: I0215 21:33:16.868028 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:26 crc kubenswrapper[4735]: I0215 21:33:26.947274 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:27 crc kubenswrapper[4735]: I0215 21:33:27.002697 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cdqzr"] Feb 15 21:33:27 crc kubenswrapper[4735]: I0215 21:33:27.954193 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cdqzr" podUID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerName="registry-server" containerID="cri-o://22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df" gracePeriod=2 Feb 15 21:33:28 crc kubenswrapper[4735]: I0215 21:33:28.896422 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:28 crc kubenswrapper[4735]: I0215 21:33:28.964075 4735 generic.go:334] "Generic (PLEG): container finished" podID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerID="22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df" exitCode=0 Feb 15 21:33:28 crc kubenswrapper[4735]: I0215 21:33:28.964119 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqzr" event={"ID":"63e0cc56-a704-4b36-a0dd-628b41f05569","Type":"ContainerDied","Data":"22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df"} Feb 15 21:33:28 crc kubenswrapper[4735]: I0215 21:33:28.964152 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqzr" event={"ID":"63e0cc56-a704-4b36-a0dd-628b41f05569","Type":"ContainerDied","Data":"b05234da8bce7576b7ccc91a426b8339700e46959d439985f786ae7616217c21"} Feb 15 21:33:28 crc kubenswrapper[4735]: I0215 21:33:28.964188 4735 scope.go:117] "RemoveContainer" containerID="22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df" Feb 15 21:33:28 crc kubenswrapper[4735]: I0215 21:33:28.964399 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdqzr" Feb 15 21:33:28 crc kubenswrapper[4735]: I0215 21:33:28.984332 4735 scope.go:117] "RemoveContainer" containerID="2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.002989 4735 scope.go:117] "RemoveContainer" containerID="fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.041759 4735 scope.go:117] "RemoveContainer" containerID="22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df" Feb 15 21:33:29 crc kubenswrapper[4735]: E0215 21:33:29.042211 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df\": container with ID starting with 22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df not found: ID does not exist" containerID="22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.042255 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df"} err="failed to get container status \"22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df\": rpc error: code = NotFound desc = could not find container \"22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df\": container with ID starting with 22d92d9c4e2e275a8a98bf6b372e7a5c2bfd2903af763f1c1a970036760e64df not found: ID does not exist" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.042286 4735 scope.go:117] "RemoveContainer" containerID="2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b" Feb 15 21:33:29 crc kubenswrapper[4735]: E0215 21:33:29.042576 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b\": container with ID starting with 2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b not found: ID does not exist" containerID="2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.042618 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b"} err="failed to get container status \"2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b\": rpc error: code = NotFound desc = could not find container \"2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b\": container with ID starting with 2be35dd13b4f10178daa1951492ec8439fddec8b1ca0ea8e8241112278bfc26b not found: ID does not exist" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.042642 4735 scope.go:117] "RemoveContainer" containerID="fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be" Feb 15 21:33:29 crc kubenswrapper[4735]: E0215 21:33:29.042852 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be\": container with ID starting with fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be not found: ID does not exist" containerID="fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.042877 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be"} err="failed to get container status \"fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be\": rpc error: code = NotFound desc = could not find container \"fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be\": container with ID starting with fe92df7edf6569531b44fda7ab116c58bbd1ddf284de8764dece01dfcc2172be not found: ID does not exist" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.047452 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxfk4\" (UniqueName: \"kubernetes.io/projected/63e0cc56-a704-4b36-a0dd-628b41f05569-kube-api-access-hxfk4\") pod \"63e0cc56-a704-4b36-a0dd-628b41f05569\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.047639 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-utilities\") pod \"63e0cc56-a704-4b36-a0dd-628b41f05569\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.047710 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-catalog-content\") pod \"63e0cc56-a704-4b36-a0dd-628b41f05569\" (UID: \"63e0cc56-a704-4b36-a0dd-628b41f05569\") " Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.052093 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-utilities" (OuterVolumeSpecName: "utilities") pod "63e0cc56-a704-4b36-a0dd-628b41f05569" (UID: "63e0cc56-a704-4b36-a0dd-628b41f05569"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.054887 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63e0cc56-a704-4b36-a0dd-628b41f05569-kube-api-access-hxfk4" (OuterVolumeSpecName: "kube-api-access-hxfk4") pod "63e0cc56-a704-4b36-a0dd-628b41f05569" (UID: "63e0cc56-a704-4b36-a0dd-628b41f05569"). InnerVolumeSpecName "kube-api-access-hxfk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.095429 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63e0cc56-a704-4b36-a0dd-628b41f05569" (UID: "63e0cc56-a704-4b36-a0dd-628b41f05569"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.149770 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.149807 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63e0cc56-a704-4b36-a0dd-628b41f05569-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.149819 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxfk4\" (UniqueName: \"kubernetes.io/projected/63e0cc56-a704-4b36-a0dd-628b41f05569-kube-api-access-hxfk4\") on node \"crc\" DevicePath \"\"" Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.301054 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cdqzr"] Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.310715 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cdqzr"] Feb 15 21:33:29 crc kubenswrapper[4735]: I0215 21:33:29.886496 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:33:29 crc kubenswrapper[4735]: E0215 21:33:29.887005 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:33:30 crc kubenswrapper[4735]: I0215 21:33:30.899619 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63e0cc56-a704-4b36-a0dd-628b41f05569" path="/var/lib/kubelet/pods/63e0cc56-a704-4b36-a0dd-628b41f05569/volumes" Feb 15 21:33:42 crc kubenswrapper[4735]: I0215 21:33:42.900344 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:33:42 crc kubenswrapper[4735]: E0215 21:33:42.901178 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:33:53 crc kubenswrapper[4735]: I0215 21:33:53.642444 4735 scope.go:117] "RemoveContainer" containerID="d963cf4e01a2739498b2f98fb02b10f4dc5e09e72b52c7e6f1a1c7949ff0f3c0" Feb 15 21:33:56 crc kubenswrapper[4735]: I0215 21:33:56.887809 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:33:56 crc kubenswrapper[4735]: E0215 21:33:56.888639 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:34:01 crc kubenswrapper[4735]: I0215 21:34:01.290979 4735 generic.go:334] "Generic (PLEG): container finished" podID="694be1b6-7203-4035-a89f-e254957c6311" containerID="9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991" exitCode=0 Feb 15 21:34:01 crc kubenswrapper[4735]: I0215 21:34:01.291481 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fgc7/must-gather-b78nv" event={"ID":"694be1b6-7203-4035-a89f-e254957c6311","Type":"ContainerDied","Data":"9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991"} Feb 15 21:34:01 crc kubenswrapper[4735]: I0215 21:34:01.292437 4735 scope.go:117] "RemoveContainer" containerID="9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991" Feb 15 21:34:01 crc kubenswrapper[4735]: I0215 21:34:01.965877 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4fgc7_must-gather-b78nv_694be1b6-7203-4035-a89f-e254957c6311/gather/0.log" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.017557 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-djtf8"] Feb 15 21:34:04 crc kubenswrapper[4735]: E0215 21:34:04.018055 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerName="extract-utilities" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.018073 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerName="extract-utilities" Feb 15 21:34:04 crc kubenswrapper[4735]: E0215 21:34:04.018100 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerName="registry-server" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.018108 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerName="registry-server" Feb 15 21:34:04 crc kubenswrapper[4735]: E0215 21:34:04.018132 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerName="extract-content" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.018140 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerName="extract-content" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.018391 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="63e0cc56-a704-4b36-a0dd-628b41f05569" containerName="registry-server" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.020436 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.043594 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djtf8"] Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.149167 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj48q\" (UniqueName: \"kubernetes.io/projected/c2630c19-ab96-42fe-bc7a-94e8255e7b45-kube-api-access-qj48q\") pod \"redhat-operators-djtf8\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.149237 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-catalog-content\") pod \"redhat-operators-djtf8\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.149309 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-utilities\") pod \"redhat-operators-djtf8\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.250927 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-utilities\") pod \"redhat-operators-djtf8\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.250356 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-utilities\") pod \"redhat-operators-djtf8\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.251163 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj48q\" (UniqueName: \"kubernetes.io/projected/c2630c19-ab96-42fe-bc7a-94e8255e7b45-kube-api-access-qj48q\") pod \"redhat-operators-djtf8\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.251215 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-catalog-content\") pod \"redhat-operators-djtf8\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.251504 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-catalog-content\") pod \"redhat-operators-djtf8\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.274875 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj48q\" (UniqueName: \"kubernetes.io/projected/c2630c19-ab96-42fe-bc7a-94e8255e7b45-kube-api-access-qj48q\") pod \"redhat-operators-djtf8\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.353053 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:04 crc kubenswrapper[4735]: I0215 21:34:04.877761 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djtf8"] Feb 15 21:34:04 crc kubenswrapper[4735]: W0215 21:34:04.889086 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2630c19_ab96_42fe_bc7a_94e8255e7b45.slice/crio-22d4cd08a58a5a92dea4a5656ed71d797242b9e8808e71832c9b45af52ffe6a4 WatchSource:0}: Error finding container 22d4cd08a58a5a92dea4a5656ed71d797242b9e8808e71832c9b45af52ffe6a4: Status 404 returned error can't find the container with id 22d4cd08a58a5a92dea4a5656ed71d797242b9e8808e71832c9b45af52ffe6a4 Feb 15 21:34:05 crc kubenswrapper[4735]: I0215 21:34:05.349132 4735 generic.go:334] "Generic (PLEG): container finished" podID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerID="1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb" exitCode=0 Feb 15 21:34:05 crc kubenswrapper[4735]: I0215 21:34:05.349234 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djtf8" event={"ID":"c2630c19-ab96-42fe-bc7a-94e8255e7b45","Type":"ContainerDied","Data":"1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb"} Feb 15 21:34:05 crc kubenswrapper[4735]: I0215 21:34:05.349478 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djtf8" event={"ID":"c2630c19-ab96-42fe-bc7a-94e8255e7b45","Type":"ContainerStarted","Data":"22d4cd08a58a5a92dea4a5656ed71d797242b9e8808e71832c9b45af52ffe6a4"} Feb 15 21:34:06 crc kubenswrapper[4735]: I0215 21:34:06.363174 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djtf8" event={"ID":"c2630c19-ab96-42fe-bc7a-94e8255e7b45","Type":"ContainerStarted","Data":"1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702"} Feb 15 21:34:08 crc kubenswrapper[4735]: I0215 21:34:08.887052 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:34:08 crc kubenswrapper[4735]: E0215 21:34:08.887876 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:34:10 crc kubenswrapper[4735]: I0215 21:34:10.656806 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4fgc7/must-gather-b78nv"] Feb 15 21:34:10 crc kubenswrapper[4735]: I0215 21:34:10.657378 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4fgc7/must-gather-b78nv" podUID="694be1b6-7203-4035-a89f-e254957c6311" containerName="copy" containerID="cri-o://bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734" gracePeriod=2 Feb 15 21:34:10 crc kubenswrapper[4735]: I0215 21:34:10.666582 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4fgc7/must-gather-b78nv"] Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.243345 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4fgc7_must-gather-b78nv_694be1b6-7203-4035-a89f-e254957c6311/copy/0.log" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.244150 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.321798 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rt2c\" (UniqueName: \"kubernetes.io/projected/694be1b6-7203-4035-a89f-e254957c6311-kube-api-access-2rt2c\") pod \"694be1b6-7203-4035-a89f-e254957c6311\" (UID: \"694be1b6-7203-4035-a89f-e254957c6311\") " Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.321935 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/694be1b6-7203-4035-a89f-e254957c6311-must-gather-output\") pod \"694be1b6-7203-4035-a89f-e254957c6311\" (UID: \"694be1b6-7203-4035-a89f-e254957c6311\") " Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.341978 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/694be1b6-7203-4035-a89f-e254957c6311-kube-api-access-2rt2c" (OuterVolumeSpecName: "kube-api-access-2rt2c") pod "694be1b6-7203-4035-a89f-e254957c6311" (UID: "694be1b6-7203-4035-a89f-e254957c6311"). InnerVolumeSpecName "kube-api-access-2rt2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.425374 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rt2c\" (UniqueName: \"kubernetes.io/projected/694be1b6-7203-4035-a89f-e254957c6311-kube-api-access-2rt2c\") on node \"crc\" DevicePath \"\"" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.430345 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4fgc7_must-gather-b78nv_694be1b6-7203-4035-a89f-e254957c6311/copy/0.log" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.431161 4735 generic.go:334] "Generic (PLEG): container finished" podID="694be1b6-7203-4035-a89f-e254957c6311" containerID="bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734" exitCode=143 Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.431212 4735 scope.go:117] "RemoveContainer" containerID="bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.431332 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fgc7/must-gather-b78nv" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.486817 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/694be1b6-7203-4035-a89f-e254957c6311-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "694be1b6-7203-4035-a89f-e254957c6311" (UID: "694be1b6-7203-4035-a89f-e254957c6311"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.492721 4735 scope.go:117] "RemoveContainer" containerID="9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.526970 4735 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/694be1b6-7203-4035-a89f-e254957c6311-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.577871 4735 scope.go:117] "RemoveContainer" containerID="bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734" Feb 15 21:34:11 crc kubenswrapper[4735]: E0215 21:34:11.579204 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734\": container with ID starting with bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734 not found: ID does not exist" containerID="bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.579241 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734"} err="failed to get container status \"bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734\": rpc error: code = NotFound desc = could not find container \"bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734\": container with ID starting with bd388db644185c2bff25c2f51249eefc619d85a04261cc7aa70c9d489b2e9734 not found: ID does not exist" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.579259 4735 scope.go:117] "RemoveContainer" containerID="9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991" Feb 15 21:34:11 crc kubenswrapper[4735]: E0215 21:34:11.579710 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991\": container with ID starting with 9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991 not found: ID does not exist" containerID="9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991" Feb 15 21:34:11 crc kubenswrapper[4735]: I0215 21:34:11.579759 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991"} err="failed to get container status \"9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991\": rpc error: code = NotFound desc = could not find container \"9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991\": container with ID starting with 9942f688247ba708c5af27e5483deed629542e2182d93c7e2b7d229011f9b991 not found: ID does not exist" Feb 15 21:34:12 crc kubenswrapper[4735]: I0215 21:34:12.442790 4735 generic.go:334] "Generic (PLEG): container finished" podID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerID="1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702" exitCode=0 Feb 15 21:34:12 crc kubenswrapper[4735]: I0215 21:34:12.442863 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djtf8" event={"ID":"c2630c19-ab96-42fe-bc7a-94e8255e7b45","Type":"ContainerDied","Data":"1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702"} Feb 15 21:34:12 crc kubenswrapper[4735]: I0215 21:34:12.913307 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="694be1b6-7203-4035-a89f-e254957c6311" path="/var/lib/kubelet/pods/694be1b6-7203-4035-a89f-e254957c6311/volumes" Feb 15 21:34:13 crc kubenswrapper[4735]: I0215 21:34:13.455869 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djtf8" event={"ID":"c2630c19-ab96-42fe-bc7a-94e8255e7b45","Type":"ContainerStarted","Data":"724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2"} Feb 15 21:34:13 crc kubenswrapper[4735]: I0215 21:34:13.485089 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-djtf8" podStartSLOduration=2.982331624 podStartE2EDuration="10.485066792s" podCreationTimestamp="2026-02-15 21:34:03 +0000 UTC" firstStartedPulling="2026-02-15 21:34:05.350797735 +0000 UTC m=+4653.216813358" lastFinishedPulling="2026-02-15 21:34:12.853532893 +0000 UTC m=+4660.719548526" observedRunningTime="2026-02-15 21:34:13.475251268 +0000 UTC m=+4661.341266901" watchObservedRunningTime="2026-02-15 21:34:13.485066792 +0000 UTC m=+4661.351082415" Feb 15 21:34:14 crc kubenswrapper[4735]: I0215 21:34:14.353849 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:14 crc kubenswrapper[4735]: I0215 21:34:14.353903 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:15 crc kubenswrapper[4735]: I0215 21:34:15.412324 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-djtf8" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerName="registry-server" probeResult="failure" output=< Feb 15 21:34:15 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:34:15 crc kubenswrapper[4735]: > Feb 15 21:34:21 crc kubenswrapper[4735]: I0215 21:34:21.886809 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:34:21 crc kubenswrapper[4735]: E0215 21:34:21.887596 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:34:24 crc kubenswrapper[4735]: I0215 21:34:24.422650 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:24 crc kubenswrapper[4735]: I0215 21:34:24.504984 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:24 crc kubenswrapper[4735]: I0215 21:34:24.664055 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djtf8"] Feb 15 21:34:25 crc kubenswrapper[4735]: I0215 21:34:25.647009 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-djtf8" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerName="registry-server" containerID="cri-o://724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2" gracePeriod=2 Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.143255 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.250178 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj48q\" (UniqueName: \"kubernetes.io/projected/c2630c19-ab96-42fe-bc7a-94e8255e7b45-kube-api-access-qj48q\") pod \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.250291 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-catalog-content\") pod \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.250423 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-utilities\") pod \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\" (UID: \"c2630c19-ab96-42fe-bc7a-94e8255e7b45\") " Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.251373 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-utilities" (OuterVolumeSpecName: "utilities") pod "c2630c19-ab96-42fe-bc7a-94e8255e7b45" (UID: "c2630c19-ab96-42fe-bc7a-94e8255e7b45"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.265144 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2630c19-ab96-42fe-bc7a-94e8255e7b45-kube-api-access-qj48q" (OuterVolumeSpecName: "kube-api-access-qj48q") pod "c2630c19-ab96-42fe-bc7a-94e8255e7b45" (UID: "c2630c19-ab96-42fe-bc7a-94e8255e7b45"). InnerVolumeSpecName "kube-api-access-qj48q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.352989 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.353038 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj48q\" (UniqueName: \"kubernetes.io/projected/c2630c19-ab96-42fe-bc7a-94e8255e7b45-kube-api-access-qj48q\") on node \"crc\" DevicePath \"\"" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.387357 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2630c19-ab96-42fe-bc7a-94e8255e7b45" (UID: "c2630c19-ab96-42fe-bc7a-94e8255e7b45"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.455138 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2630c19-ab96-42fe-bc7a-94e8255e7b45-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.656913 4735 generic.go:334] "Generic (PLEG): container finished" podID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerID="724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2" exitCode=0 Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.656993 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djtf8" event={"ID":"c2630c19-ab96-42fe-bc7a-94e8255e7b45","Type":"ContainerDied","Data":"724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2"} Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.657199 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djtf8" event={"ID":"c2630c19-ab96-42fe-bc7a-94e8255e7b45","Type":"ContainerDied","Data":"22d4cd08a58a5a92dea4a5656ed71d797242b9e8808e71832c9b45af52ffe6a4"} Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.657220 4735 scope.go:117] "RemoveContainer" containerID="724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.657045 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djtf8" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.696236 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djtf8"] Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.697397 4735 scope.go:117] "RemoveContainer" containerID="1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.709182 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-djtf8"] Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.730878 4735 scope.go:117] "RemoveContainer" containerID="1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.767338 4735 scope.go:117] "RemoveContainer" containerID="724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2" Feb 15 21:34:26 crc kubenswrapper[4735]: E0215 21:34:26.769597 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2\": container with ID starting with 724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2 not found: ID does not exist" containerID="724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.769688 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2"} err="failed to get container status \"724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2\": rpc error: code = NotFound desc = could not find container \"724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2\": container with ID starting with 724916f90bc42cce155107e1e2f996477887495c88875afebdd389dde73e0dd2 not found: ID does not exist" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.769762 4735 scope.go:117] "RemoveContainer" containerID="1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702" Feb 15 21:34:26 crc kubenswrapper[4735]: E0215 21:34:26.770297 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702\": container with ID starting with 1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702 not found: ID does not exist" containerID="1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.770343 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702"} err="failed to get container status \"1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702\": rpc error: code = NotFound desc = could not find container \"1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702\": container with ID starting with 1d397dca940cb131e4594d5c45bb9c09951bd22c5b9bdc589bf09f63ae83c702 not found: ID does not exist" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.770360 4735 scope.go:117] "RemoveContainer" containerID="1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb" Feb 15 21:34:26 crc kubenswrapper[4735]: E0215 21:34:26.770609 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb\": container with ID starting with 1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb not found: ID does not exist" containerID="1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.770633 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb"} err="failed to get container status \"1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb\": rpc error: code = NotFound desc = could not find container \"1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb\": container with ID starting with 1b57f308bc1ca7ee2da8765b66ce709dcb2b8871239bbade26c18f430f189ebb not found: ID does not exist" Feb 15 21:34:26 crc kubenswrapper[4735]: I0215 21:34:26.896736 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" path="/var/lib/kubelet/pods/c2630c19-ab96-42fe-bc7a-94e8255e7b45/volumes" Feb 15 21:34:32 crc kubenswrapper[4735]: I0215 21:34:32.897188 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:34:32 crc kubenswrapper[4735]: E0215 21:34:32.898077 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:34:43 crc kubenswrapper[4735]: I0215 21:34:43.888686 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:34:43 crc kubenswrapper[4735]: E0215 21:34:43.891872 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:34:53 crc kubenswrapper[4735]: I0215 21:34:53.732621 4735 scope.go:117] "RemoveContainer" containerID="07391ba50ab8f8b18b852f5e536b0e09724bfe790306f200ddd8bc8cef71ea58" Feb 15 21:34:54 crc kubenswrapper[4735]: I0215 21:34:54.887054 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:34:54 crc kubenswrapper[4735]: E0215 21:34:54.887683 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:35:06 crc kubenswrapper[4735]: I0215 21:35:06.887483 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:35:06 crc kubenswrapper[4735]: E0215 21:35:06.888992 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:35:17 crc kubenswrapper[4735]: I0215 21:35:17.455704 4735 patch_prober.go:28] interesting pod/router-default-5444994796-lj984 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 21:35:17 crc kubenswrapper[4735]: I0215 21:35:17.456549 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-lj984" podUID="d699b610-3540-4383-9af6-257c6821c498" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 15 21:35:17 crc kubenswrapper[4735]: I0215 21:35:17.457337 4735 patch_prober.go:28] interesting pod/nmstate-webhook-866bcb46dc-7gkdk container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.32:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 15 21:35:17 crc kubenswrapper[4735]: I0215 21:35:17.457384 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-7gkdk" podUID="cecb1f2e-e07b-4ea1-924d-7802fa40fd6b" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.32:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 15 21:35:17 crc kubenswrapper[4735]: I0215 21:35:17.463243 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-ldbt6" podUID="32c1d680-92cf-4c49-93b9-17bb17aaf623" containerName="registry-server" probeResult="failure" output=< Feb 15 21:35:17 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:35:17 crc kubenswrapper[4735]: > Feb 15 21:35:17 crc kubenswrapper[4735]: I0215 21:35:17.472509 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-ldbt6" podUID="32c1d680-92cf-4c49-93b9-17bb17aaf623" containerName="registry-server" probeResult="failure" output=< Feb 15 21:35:17 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:35:17 crc kubenswrapper[4735]: > Feb 15 21:35:18 crc kubenswrapper[4735]: I0215 21:35:18.886802 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:35:18 crc kubenswrapper[4735]: E0215 21:35:18.887506 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:35:33 crc kubenswrapper[4735]: I0215 21:35:33.886913 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:35:33 crc kubenswrapper[4735]: E0215 21:35:33.888026 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:35:44 crc kubenswrapper[4735]: I0215 21:35:44.886850 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:35:44 crc kubenswrapper[4735]: E0215 21:35:44.891179 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:35:57 crc kubenswrapper[4735]: I0215 21:35:57.887638 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:35:59 crc kubenswrapper[4735]: I0215 21:35:59.551788 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"79803829e7f502800603cb50ee8fbc1de3db60bf3e331d273572d882ec3710ef"} Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.358020 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-667js"] Feb 15 21:36:47 crc kubenswrapper[4735]: E0215 21:36:47.358992 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerName="registry-server" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.359006 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerName="registry-server" Feb 15 21:36:47 crc kubenswrapper[4735]: E0215 21:36:47.359026 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="694be1b6-7203-4035-a89f-e254957c6311" containerName="copy" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.359032 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="694be1b6-7203-4035-a89f-e254957c6311" containerName="copy" Feb 15 21:36:47 crc kubenswrapper[4735]: E0215 21:36:47.359049 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerName="extract-utilities" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.359055 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerName="extract-utilities" Feb 15 21:36:47 crc kubenswrapper[4735]: E0215 21:36:47.359071 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="694be1b6-7203-4035-a89f-e254957c6311" containerName="gather" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.359077 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="694be1b6-7203-4035-a89f-e254957c6311" containerName="gather" Feb 15 21:36:47 crc kubenswrapper[4735]: E0215 21:36:47.359089 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerName="extract-content" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.359096 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerName="extract-content" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.359257 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="694be1b6-7203-4035-a89f-e254957c6311" containerName="gather" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.359277 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="694be1b6-7203-4035-a89f-e254957c6311" containerName="copy" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.359284 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2630c19-ab96-42fe-bc7a-94e8255e7b45" containerName="registry-server" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.363126 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.408411 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-667js"] Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.420762 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4ftc\" (UniqueName: \"kubernetes.io/projected/230dbf02-58ab-4090-ad8b-85a1d519d6bf-kube-api-access-h4ftc\") pod \"redhat-marketplace-667js\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.420904 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-utilities\") pod \"redhat-marketplace-667js\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.421002 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-catalog-content\") pod \"redhat-marketplace-667js\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.522227 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-utilities\") pod \"redhat-marketplace-667js\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.522278 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-catalog-content\") pod \"redhat-marketplace-667js\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.522377 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4ftc\" (UniqueName: \"kubernetes.io/projected/230dbf02-58ab-4090-ad8b-85a1d519d6bf-kube-api-access-h4ftc\") pod \"redhat-marketplace-667js\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.523162 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-utilities\") pod \"redhat-marketplace-667js\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.523375 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-catalog-content\") pod \"redhat-marketplace-667js\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.547106 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4ftc\" (UniqueName: \"kubernetes.io/projected/230dbf02-58ab-4090-ad8b-85a1d519d6bf-kube-api-access-h4ftc\") pod \"redhat-marketplace-667js\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:47 crc kubenswrapper[4735]: I0215 21:36:47.688648 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:48 crc kubenswrapper[4735]: I0215 21:36:48.196889 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-667js"] Feb 15 21:36:49 crc kubenswrapper[4735]: I0215 21:36:49.048274 4735 generic.go:334] "Generic (PLEG): container finished" podID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerID="6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59" exitCode=0 Feb 15 21:36:49 crc kubenswrapper[4735]: I0215 21:36:49.048366 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667js" event={"ID":"230dbf02-58ab-4090-ad8b-85a1d519d6bf","Type":"ContainerDied","Data":"6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59"} Feb 15 21:36:49 crc kubenswrapper[4735]: I0215 21:36:49.048596 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667js" event={"ID":"230dbf02-58ab-4090-ad8b-85a1d519d6bf","Type":"ContainerStarted","Data":"43f66b6f36b58a17caf8cf17e6cdbd4a37b457b6d3d2f7d206681dfb7aa626f7"} Feb 15 21:36:50 crc kubenswrapper[4735]: I0215 21:36:50.063282 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667js" event={"ID":"230dbf02-58ab-4090-ad8b-85a1d519d6bf","Type":"ContainerStarted","Data":"db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc"} Feb 15 21:36:51 crc kubenswrapper[4735]: I0215 21:36:51.071342 4735 generic.go:334] "Generic (PLEG): container finished" podID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerID="db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc" exitCode=0 Feb 15 21:36:51 crc kubenswrapper[4735]: I0215 21:36:51.071386 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667js" event={"ID":"230dbf02-58ab-4090-ad8b-85a1d519d6bf","Type":"ContainerDied","Data":"db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc"} Feb 15 21:36:52 crc kubenswrapper[4735]: I0215 21:36:52.082172 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667js" event={"ID":"230dbf02-58ab-4090-ad8b-85a1d519d6bf","Type":"ContainerStarted","Data":"fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7"} Feb 15 21:36:52 crc kubenswrapper[4735]: I0215 21:36:52.099246 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-667js" podStartSLOduration=2.4474957059999998 podStartE2EDuration="5.099225887s" podCreationTimestamp="2026-02-15 21:36:47 +0000 UTC" firstStartedPulling="2026-02-15 21:36:49.052526446 +0000 UTC m=+4816.918542059" lastFinishedPulling="2026-02-15 21:36:51.704256617 +0000 UTC m=+4819.570272240" observedRunningTime="2026-02-15 21:36:52.096981617 +0000 UTC m=+4819.962997230" watchObservedRunningTime="2026-02-15 21:36:52.099225887 +0000 UTC m=+4819.965241520" Feb 15 21:36:57 crc kubenswrapper[4735]: I0215 21:36:57.689616 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:57 crc kubenswrapper[4735]: I0215 21:36:57.690098 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:57 crc kubenswrapper[4735]: I0215 21:36:57.759781 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:58 crc kubenswrapper[4735]: I0215 21:36:58.201105 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:36:58 crc kubenswrapper[4735]: I0215 21:36:58.255593 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-667js"] Feb 15 21:37:00 crc kubenswrapper[4735]: I0215 21:37:00.151747 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-667js" podUID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerName="registry-server" containerID="cri-o://fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7" gracePeriod=2 Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.126667 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.163618 4735 generic.go:334] "Generic (PLEG): container finished" podID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerID="fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7" exitCode=0 Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.163670 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667js" event={"ID":"230dbf02-58ab-4090-ad8b-85a1d519d6bf","Type":"ContainerDied","Data":"fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7"} Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.163705 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-667js" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.163732 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667js" event={"ID":"230dbf02-58ab-4090-ad8b-85a1d519d6bf","Type":"ContainerDied","Data":"43f66b6f36b58a17caf8cf17e6cdbd4a37b457b6d3d2f7d206681dfb7aa626f7"} Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.163776 4735 scope.go:117] "RemoveContainer" containerID="fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.185220 4735 scope.go:117] "RemoveContainer" containerID="db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.206400 4735 scope.go:117] "RemoveContainer" containerID="6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.244909 4735 scope.go:117] "RemoveContainer" containerID="fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7" Feb 15 21:37:01 crc kubenswrapper[4735]: E0215 21:37:01.245330 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7\": container with ID starting with fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7 not found: ID does not exist" containerID="fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.245363 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7"} err="failed to get container status \"fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7\": rpc error: code = NotFound desc = could not find container \"fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7\": container with ID starting with fc2a863a27c52bb515a139f555ea324eacec9957c8b51cf7258c660a8bd4caa7 not found: ID does not exist" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.245383 4735 scope.go:117] "RemoveContainer" containerID="db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc" Feb 15 21:37:01 crc kubenswrapper[4735]: E0215 21:37:01.245615 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc\": container with ID starting with db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc not found: ID does not exist" containerID="db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.245637 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc"} err="failed to get container status \"db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc\": rpc error: code = NotFound desc = could not find container \"db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc\": container with ID starting with db3c492361e078c9694be54e1aff50f9a9eac9fe347d150cfc5ff5e1f5cd7bcc not found: ID does not exist" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.245649 4735 scope.go:117] "RemoveContainer" containerID="6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59" Feb 15 21:37:01 crc kubenswrapper[4735]: E0215 21:37:01.245889 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59\": container with ID starting with 6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59 not found: ID does not exist" containerID="6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.245910 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59"} err="failed to get container status \"6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59\": rpc error: code = NotFound desc = could not find container \"6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59\": container with ID starting with 6070122605ec8435bb4292ee53cf46f45233f768bb2df07f9ea376e9d6a53b59 not found: ID does not exist" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.316234 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-catalog-content\") pod \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.316711 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4ftc\" (UniqueName: \"kubernetes.io/projected/230dbf02-58ab-4090-ad8b-85a1d519d6bf-kube-api-access-h4ftc\") pod \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.316901 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-utilities\") pod \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\" (UID: \"230dbf02-58ab-4090-ad8b-85a1d519d6bf\") " Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.317808 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-utilities" (OuterVolumeSpecName: "utilities") pod "230dbf02-58ab-4090-ad8b-85a1d519d6bf" (UID: "230dbf02-58ab-4090-ad8b-85a1d519d6bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.326113 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/230dbf02-58ab-4090-ad8b-85a1d519d6bf-kube-api-access-h4ftc" (OuterVolumeSpecName: "kube-api-access-h4ftc") pod "230dbf02-58ab-4090-ad8b-85a1d519d6bf" (UID: "230dbf02-58ab-4090-ad8b-85a1d519d6bf"). InnerVolumeSpecName "kube-api-access-h4ftc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.356687 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "230dbf02-58ab-4090-ad8b-85a1d519d6bf" (UID: "230dbf02-58ab-4090-ad8b-85a1d519d6bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.419685 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.419911 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4ftc\" (UniqueName: \"kubernetes.io/projected/230dbf02-58ab-4090-ad8b-85a1d519d6bf-kube-api-access-h4ftc\") on node \"crc\" DevicePath \"\"" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.420084 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbf02-58ab-4090-ad8b-85a1d519d6bf-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.510197 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-667js"] Feb 15 21:37:01 crc kubenswrapper[4735]: I0215 21:37:01.520527 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-667js"] Feb 15 21:37:02 crc kubenswrapper[4735]: I0215 21:37:02.903584 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" path="/var/lib/kubelet/pods/230dbf02-58ab-4090-ad8b-85a1d519d6bf/volumes" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.587750 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tvx7p/must-gather-7657q"] Feb 15 21:37:18 crc kubenswrapper[4735]: E0215 21:37:18.588732 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerName="registry-server" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.588749 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerName="registry-server" Feb 15 21:37:18 crc kubenswrapper[4735]: E0215 21:37:18.588764 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerName="extract-utilities" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.588772 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerName="extract-utilities" Feb 15 21:37:18 crc kubenswrapper[4735]: E0215 21:37:18.588809 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerName="extract-content" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.588816 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerName="extract-content" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.597158 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="230dbf02-58ab-4090-ad8b-85a1d519d6bf" containerName="registry-server" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.598256 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.605892 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tvx7p"/"kube-root-ca.crt" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.605998 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tvx7p"/"openshift-service-ca.crt" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.673103 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2hn7\" (UniqueName: \"kubernetes.io/projected/f92d063b-e0fa-401e-9a93-30fadc3e2425-kube-api-access-c2hn7\") pod \"must-gather-7657q\" (UID: \"f92d063b-e0fa-401e-9a93-30fadc3e2425\") " pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.673177 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f92d063b-e0fa-401e-9a93-30fadc3e2425-must-gather-output\") pod \"must-gather-7657q\" (UID: \"f92d063b-e0fa-401e-9a93-30fadc3e2425\") " pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.719661 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tvx7p/must-gather-7657q"] Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.774592 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2hn7\" (UniqueName: \"kubernetes.io/projected/f92d063b-e0fa-401e-9a93-30fadc3e2425-kube-api-access-c2hn7\") pod \"must-gather-7657q\" (UID: \"f92d063b-e0fa-401e-9a93-30fadc3e2425\") " pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.774662 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f92d063b-e0fa-401e-9a93-30fadc3e2425-must-gather-output\") pod \"must-gather-7657q\" (UID: \"f92d063b-e0fa-401e-9a93-30fadc3e2425\") " pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.775140 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f92d063b-e0fa-401e-9a93-30fadc3e2425-must-gather-output\") pod \"must-gather-7657q\" (UID: \"f92d063b-e0fa-401e-9a93-30fadc3e2425\") " pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.806645 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2hn7\" (UniqueName: \"kubernetes.io/projected/f92d063b-e0fa-401e-9a93-30fadc3e2425-kube-api-access-c2hn7\") pod \"must-gather-7657q\" (UID: \"f92d063b-e0fa-401e-9a93-30fadc3e2425\") " pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:37:18 crc kubenswrapper[4735]: I0215 21:37:18.924151 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:37:19 crc kubenswrapper[4735]: I0215 21:37:19.456384 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tvx7p/must-gather-7657q"] Feb 15 21:37:20 crc kubenswrapper[4735]: I0215 21:37:20.411900 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/must-gather-7657q" event={"ID":"f92d063b-e0fa-401e-9a93-30fadc3e2425","Type":"ContainerStarted","Data":"81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d"} Feb 15 21:37:20 crc kubenswrapper[4735]: I0215 21:37:20.412539 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/must-gather-7657q" event={"ID":"f92d063b-e0fa-401e-9a93-30fadc3e2425","Type":"ContainerStarted","Data":"bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b"} Feb 15 21:37:20 crc kubenswrapper[4735]: I0215 21:37:20.412556 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/must-gather-7657q" event={"ID":"f92d063b-e0fa-401e-9a93-30fadc3e2425","Type":"ContainerStarted","Data":"c3dca49a82a4ef913e3a40cba4c3d6f3911609a583d7a930ebc6e94b0061b223"} Feb 15 21:37:20 crc kubenswrapper[4735]: I0215 21:37:20.435197 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tvx7p/must-gather-7657q" podStartSLOduration=2.435170077 podStartE2EDuration="2.435170077s" podCreationTimestamp="2026-02-15 21:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 21:37:20.424693096 +0000 UTC m=+4848.290708719" watchObservedRunningTime="2026-02-15 21:37:20.435170077 +0000 UTC m=+4848.301185730" Feb 15 21:37:23 crc kubenswrapper[4735]: I0215 21:37:23.764055 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tvx7p/crc-debug-zx994"] Feb 15 21:37:23 crc kubenswrapper[4735]: I0215 21:37:23.765721 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:37:23 crc kubenswrapper[4735]: I0215 21:37:23.768200 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-tvx7p"/"default-dockercfg-h2zgf" Feb 15 21:37:23 crc kubenswrapper[4735]: I0215 21:37:23.891046 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf92d\" (UniqueName: \"kubernetes.io/projected/ab728743-67c2-48f8-a242-9cbb485e130f-kube-api-access-tf92d\") pod \"crc-debug-zx994\" (UID: \"ab728743-67c2-48f8-a242-9cbb485e130f\") " pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:37:23 crc kubenswrapper[4735]: I0215 21:37:23.891391 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab728743-67c2-48f8-a242-9cbb485e130f-host\") pod \"crc-debug-zx994\" (UID: \"ab728743-67c2-48f8-a242-9cbb485e130f\") " pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:37:23 crc kubenswrapper[4735]: I0215 21:37:23.993192 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf92d\" (UniqueName: \"kubernetes.io/projected/ab728743-67c2-48f8-a242-9cbb485e130f-kube-api-access-tf92d\") pod \"crc-debug-zx994\" (UID: \"ab728743-67c2-48f8-a242-9cbb485e130f\") " pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:37:23 crc kubenswrapper[4735]: I0215 21:37:23.993335 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab728743-67c2-48f8-a242-9cbb485e130f-host\") pod \"crc-debug-zx994\" (UID: \"ab728743-67c2-48f8-a242-9cbb485e130f\") " pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:37:23 crc kubenswrapper[4735]: I0215 21:37:23.993577 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab728743-67c2-48f8-a242-9cbb485e130f-host\") pod \"crc-debug-zx994\" (UID: \"ab728743-67c2-48f8-a242-9cbb485e130f\") " pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:37:24 crc kubenswrapper[4735]: I0215 21:37:24.013203 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf92d\" (UniqueName: \"kubernetes.io/projected/ab728743-67c2-48f8-a242-9cbb485e130f-kube-api-access-tf92d\") pod \"crc-debug-zx994\" (UID: \"ab728743-67c2-48f8-a242-9cbb485e130f\") " pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:37:24 crc kubenswrapper[4735]: I0215 21:37:24.089559 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:37:24 crc kubenswrapper[4735]: W0215 21:37:24.132583 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab728743_67c2_48f8_a242_9cbb485e130f.slice/crio-f030fa2cb1080a06c6ef4b9081f3d759986e1323fc6fcb76edf5841429f99b2c WatchSource:0}: Error finding container f030fa2cb1080a06c6ef4b9081f3d759986e1323fc6fcb76edf5841429f99b2c: Status 404 returned error can't find the container with id f030fa2cb1080a06c6ef4b9081f3d759986e1323fc6fcb76edf5841429f99b2c Feb 15 21:37:24 crc kubenswrapper[4735]: I0215 21:37:24.442719 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/crc-debug-zx994" event={"ID":"ab728743-67c2-48f8-a242-9cbb485e130f","Type":"ContainerStarted","Data":"768e0f78a548966702f5215b4961fb1e735a5cffea5d303c4887b6ae0ae4107b"} Feb 15 21:37:24 crc kubenswrapper[4735]: I0215 21:37:24.442902 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/crc-debug-zx994" event={"ID":"ab728743-67c2-48f8-a242-9cbb485e130f","Type":"ContainerStarted","Data":"f030fa2cb1080a06c6ef4b9081f3d759986e1323fc6fcb76edf5841429f99b2c"} Feb 15 21:37:24 crc kubenswrapper[4735]: I0215 21:37:24.469879 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tvx7p/crc-debug-zx994" podStartSLOduration=1.469856683 podStartE2EDuration="1.469856683s" podCreationTimestamp="2026-02-15 21:37:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 21:37:24.463446391 +0000 UTC m=+4852.329462014" watchObservedRunningTime="2026-02-15 21:37:24.469856683 +0000 UTC m=+4852.335872306" Feb 15 21:38:03 crc kubenswrapper[4735]: I0215 21:38:03.786003 4735 generic.go:334] "Generic (PLEG): container finished" podID="ab728743-67c2-48f8-a242-9cbb485e130f" containerID="768e0f78a548966702f5215b4961fb1e735a5cffea5d303c4887b6ae0ae4107b" exitCode=0 Feb 15 21:38:03 crc kubenswrapper[4735]: I0215 21:38:03.786070 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/crc-debug-zx994" event={"ID":"ab728743-67c2-48f8-a242-9cbb485e130f","Type":"ContainerDied","Data":"768e0f78a548966702f5215b4961fb1e735a5cffea5d303c4887b6ae0ae4107b"} Feb 15 21:38:04 crc kubenswrapper[4735]: I0215 21:38:04.903105 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:38:04 crc kubenswrapper[4735]: I0215 21:38:04.936911 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tvx7p/crc-debug-zx994"] Feb 15 21:38:04 crc kubenswrapper[4735]: I0215 21:38:04.946220 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tvx7p/crc-debug-zx994"] Feb 15 21:38:04 crc kubenswrapper[4735]: I0215 21:38:04.996573 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf92d\" (UniqueName: \"kubernetes.io/projected/ab728743-67c2-48f8-a242-9cbb485e130f-kube-api-access-tf92d\") pod \"ab728743-67c2-48f8-a242-9cbb485e130f\" (UID: \"ab728743-67c2-48f8-a242-9cbb485e130f\") " Feb 15 21:38:04 crc kubenswrapper[4735]: I0215 21:38:04.996661 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab728743-67c2-48f8-a242-9cbb485e130f-host\") pod \"ab728743-67c2-48f8-a242-9cbb485e130f\" (UID: \"ab728743-67c2-48f8-a242-9cbb485e130f\") " Feb 15 21:38:04 crc kubenswrapper[4735]: I0215 21:38:04.997858 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab728743-67c2-48f8-a242-9cbb485e130f-host" (OuterVolumeSpecName: "host") pod "ab728743-67c2-48f8-a242-9cbb485e130f" (UID: "ab728743-67c2-48f8-a242-9cbb485e130f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 21:38:05 crc kubenswrapper[4735]: I0215 21:38:05.006792 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab728743-67c2-48f8-a242-9cbb485e130f-kube-api-access-tf92d" (OuterVolumeSpecName: "kube-api-access-tf92d") pod "ab728743-67c2-48f8-a242-9cbb485e130f" (UID: "ab728743-67c2-48f8-a242-9cbb485e130f"). InnerVolumeSpecName "kube-api-access-tf92d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:38:05 crc kubenswrapper[4735]: I0215 21:38:05.099011 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf92d\" (UniqueName: \"kubernetes.io/projected/ab728743-67c2-48f8-a242-9cbb485e130f-kube-api-access-tf92d\") on node \"crc\" DevicePath \"\"" Feb 15 21:38:05 crc kubenswrapper[4735]: I0215 21:38:05.099049 4735 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab728743-67c2-48f8-a242-9cbb485e130f-host\") on node \"crc\" DevicePath \"\"" Feb 15 21:38:05 crc kubenswrapper[4735]: I0215 21:38:05.807799 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f030fa2cb1080a06c6ef4b9081f3d759986e1323fc6fcb76edf5841429f99b2c" Feb 15 21:38:05 crc kubenswrapper[4735]: I0215 21:38:05.807883 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-zx994" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.300195 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tvx7p/crc-debug-zxx97"] Feb 15 21:38:06 crc kubenswrapper[4735]: E0215 21:38:06.300549 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab728743-67c2-48f8-a242-9cbb485e130f" containerName="container-00" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.300561 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab728743-67c2-48f8-a242-9cbb485e130f" containerName="container-00" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.300746 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab728743-67c2-48f8-a242-9cbb485e130f" containerName="container-00" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.301503 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.304237 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-tvx7p"/"default-dockercfg-h2zgf" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.419841 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q47bv\" (UniqueName: \"kubernetes.io/projected/cae4232e-515c-4113-b2be-bf2bc57480ac-kube-api-access-q47bv\") pod \"crc-debug-zxx97\" (UID: \"cae4232e-515c-4113-b2be-bf2bc57480ac\") " pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.420055 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cae4232e-515c-4113-b2be-bf2bc57480ac-host\") pod \"crc-debug-zxx97\" (UID: \"cae4232e-515c-4113-b2be-bf2bc57480ac\") " pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.521695 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q47bv\" (UniqueName: \"kubernetes.io/projected/cae4232e-515c-4113-b2be-bf2bc57480ac-kube-api-access-q47bv\") pod \"crc-debug-zxx97\" (UID: \"cae4232e-515c-4113-b2be-bf2bc57480ac\") " pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.522170 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cae4232e-515c-4113-b2be-bf2bc57480ac-host\") pod \"crc-debug-zxx97\" (UID: \"cae4232e-515c-4113-b2be-bf2bc57480ac\") " pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.522321 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cae4232e-515c-4113-b2be-bf2bc57480ac-host\") pod \"crc-debug-zxx97\" (UID: \"cae4232e-515c-4113-b2be-bf2bc57480ac\") " pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.539348 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q47bv\" (UniqueName: \"kubernetes.io/projected/cae4232e-515c-4113-b2be-bf2bc57480ac-kube-api-access-q47bv\") pod \"crc-debug-zxx97\" (UID: \"cae4232e-515c-4113-b2be-bf2bc57480ac\") " pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.616886 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.827104 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/crc-debug-zxx97" event={"ID":"cae4232e-515c-4113-b2be-bf2bc57480ac","Type":"ContainerStarted","Data":"e5d9be7cc453fd6e916d89696007f6aad8326eff001ac5f91c5b20a405b15acf"} Feb 15 21:38:06 crc kubenswrapper[4735]: I0215 21:38:06.896906 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab728743-67c2-48f8-a242-9cbb485e130f" path="/var/lib/kubelet/pods/ab728743-67c2-48f8-a242-9cbb485e130f/volumes" Feb 15 21:38:07 crc kubenswrapper[4735]: I0215 21:38:07.835656 4735 generic.go:334] "Generic (PLEG): container finished" podID="cae4232e-515c-4113-b2be-bf2bc57480ac" containerID="07e11bb215e64d4a85fb253da1d5e41cd3f16b9bbdefe9889bbad9160fa7dbd0" exitCode=0 Feb 15 21:38:07 crc kubenswrapper[4735]: I0215 21:38:07.835734 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/crc-debug-zxx97" event={"ID":"cae4232e-515c-4113-b2be-bf2bc57480ac","Type":"ContainerDied","Data":"07e11bb215e64d4a85fb253da1d5e41cd3f16b9bbdefe9889bbad9160fa7dbd0"} Feb 15 21:38:08 crc kubenswrapper[4735]: I0215 21:38:08.983328 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.065100 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q47bv\" (UniqueName: \"kubernetes.io/projected/cae4232e-515c-4113-b2be-bf2bc57480ac-kube-api-access-q47bv\") pod \"cae4232e-515c-4113-b2be-bf2bc57480ac\" (UID: \"cae4232e-515c-4113-b2be-bf2bc57480ac\") " Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.066093 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cae4232e-515c-4113-b2be-bf2bc57480ac-host\") pod \"cae4232e-515c-4113-b2be-bf2bc57480ac\" (UID: \"cae4232e-515c-4113-b2be-bf2bc57480ac\") " Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.066493 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cae4232e-515c-4113-b2be-bf2bc57480ac-host" (OuterVolumeSpecName: "host") pod "cae4232e-515c-4113-b2be-bf2bc57480ac" (UID: "cae4232e-515c-4113-b2be-bf2bc57480ac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.072488 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cae4232e-515c-4113-b2be-bf2bc57480ac-kube-api-access-q47bv" (OuterVolumeSpecName: "kube-api-access-q47bv") pod "cae4232e-515c-4113-b2be-bf2bc57480ac" (UID: "cae4232e-515c-4113-b2be-bf2bc57480ac"). InnerVolumeSpecName "kube-api-access-q47bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.167981 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q47bv\" (UniqueName: \"kubernetes.io/projected/cae4232e-515c-4113-b2be-bf2bc57480ac-kube-api-access-q47bv\") on node \"crc\" DevicePath \"\"" Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.168266 4735 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cae4232e-515c-4113-b2be-bf2bc57480ac-host\") on node \"crc\" DevicePath \"\"" Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.821369 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tvx7p/crc-debug-zxx97"] Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.826979 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tvx7p/crc-debug-zxx97"] Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.853097 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5d9be7cc453fd6e916d89696007f6aad8326eff001ac5f91c5b20a405b15acf" Feb 15 21:38:09 crc kubenswrapper[4735]: I0215 21:38:09.853151 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-zxx97" Feb 15 21:38:10 crc kubenswrapper[4735]: I0215 21:38:10.895839 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cae4232e-515c-4113-b2be-bf2bc57480ac" path="/var/lib/kubelet/pods/cae4232e-515c-4113-b2be-bf2bc57480ac/volumes" Feb 15 21:38:10 crc kubenswrapper[4735]: I0215 21:38:10.971371 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tvx7p/crc-debug-smpqh"] Feb 15 21:38:10 crc kubenswrapper[4735]: E0215 21:38:10.971714 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae4232e-515c-4113-b2be-bf2bc57480ac" containerName="container-00" Feb 15 21:38:10 crc kubenswrapper[4735]: I0215 21:38:10.971730 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae4232e-515c-4113-b2be-bf2bc57480ac" containerName="container-00" Feb 15 21:38:10 crc kubenswrapper[4735]: I0215 21:38:10.971928 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae4232e-515c-4113-b2be-bf2bc57480ac" containerName="container-00" Feb 15 21:38:10 crc kubenswrapper[4735]: I0215 21:38:10.972514 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:10 crc kubenswrapper[4735]: I0215 21:38:10.974484 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-tvx7p"/"default-dockercfg-h2zgf" Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.011515 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86cpb\" (UniqueName: \"kubernetes.io/projected/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-kube-api-access-86cpb\") pod \"crc-debug-smpqh\" (UID: \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\") " pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.011832 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-host\") pod \"crc-debug-smpqh\" (UID: \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\") " pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.114997 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-host\") pod \"crc-debug-smpqh\" (UID: \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\") " pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.115150 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86cpb\" (UniqueName: \"kubernetes.io/projected/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-kube-api-access-86cpb\") pod \"crc-debug-smpqh\" (UID: \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\") " pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.115151 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-host\") pod \"crc-debug-smpqh\" (UID: \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\") " pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.136183 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86cpb\" (UniqueName: \"kubernetes.io/projected/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-kube-api-access-86cpb\") pod \"crc-debug-smpqh\" (UID: \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\") " pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.290960 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:11 crc kubenswrapper[4735]: W0215 21:38:11.314980 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae6f1293_c2f8_4231_ae43_8c40d3fcf99a.slice/crio-f6169e1520110e5bd10cdbf0bebbf5905b96a3368a9fa930ee324063cd6393b2 WatchSource:0}: Error finding container f6169e1520110e5bd10cdbf0bebbf5905b96a3368a9fa930ee324063cd6393b2: Status 404 returned error can't find the container with id f6169e1520110e5bd10cdbf0bebbf5905b96a3368a9fa930ee324063cd6393b2 Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.870453 4735 generic.go:334] "Generic (PLEG): container finished" podID="ae6f1293-c2f8-4231-ae43-8c40d3fcf99a" containerID="67ef80a01879dd8096048e6d30ebdb159209bb6495db746b3e5dcdd479209143" exitCode=0 Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.870561 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/crc-debug-smpqh" event={"ID":"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a","Type":"ContainerDied","Data":"67ef80a01879dd8096048e6d30ebdb159209bb6495db746b3e5dcdd479209143"} Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.871130 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/crc-debug-smpqh" event={"ID":"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a","Type":"ContainerStarted","Data":"f6169e1520110e5bd10cdbf0bebbf5905b96a3368a9fa930ee324063cd6393b2"} Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.916748 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tvx7p/crc-debug-smpqh"] Feb 15 21:38:11 crc kubenswrapper[4735]: I0215 21:38:11.926344 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tvx7p/crc-debug-smpqh"] Feb 15 21:38:12 crc kubenswrapper[4735]: I0215 21:38:12.966853 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:13 crc kubenswrapper[4735]: I0215 21:38:13.052617 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-host\") pod \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\" (UID: \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\") " Feb 15 21:38:13 crc kubenswrapper[4735]: I0215 21:38:13.052701 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-host" (OuterVolumeSpecName: "host") pod "ae6f1293-c2f8-4231-ae43-8c40d3fcf99a" (UID: "ae6f1293-c2f8-4231-ae43-8c40d3fcf99a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 15 21:38:13 crc kubenswrapper[4735]: I0215 21:38:13.052902 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86cpb\" (UniqueName: \"kubernetes.io/projected/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-kube-api-access-86cpb\") pod \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\" (UID: \"ae6f1293-c2f8-4231-ae43-8c40d3fcf99a\") " Feb 15 21:38:13 crc kubenswrapper[4735]: I0215 21:38:13.054259 4735 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-host\") on node \"crc\" DevicePath \"\"" Feb 15 21:38:13 crc kubenswrapper[4735]: I0215 21:38:13.067430 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-kube-api-access-86cpb" (OuterVolumeSpecName: "kube-api-access-86cpb") pod "ae6f1293-c2f8-4231-ae43-8c40d3fcf99a" (UID: "ae6f1293-c2f8-4231-ae43-8c40d3fcf99a"). InnerVolumeSpecName "kube-api-access-86cpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:38:13 crc kubenswrapper[4735]: I0215 21:38:13.156059 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86cpb\" (UniqueName: \"kubernetes.io/projected/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a-kube-api-access-86cpb\") on node \"crc\" DevicePath \"\"" Feb 15 21:38:13 crc kubenswrapper[4735]: I0215 21:38:13.888721 4735 scope.go:117] "RemoveContainer" containerID="67ef80a01879dd8096048e6d30ebdb159209bb6495db746b3e5dcdd479209143" Feb 15 21:38:13 crc kubenswrapper[4735]: I0215 21:38:13.888757 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/crc-debug-smpqh" Feb 15 21:38:14 crc kubenswrapper[4735]: I0215 21:38:14.901962 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae6f1293-c2f8-4231-ae43-8c40d3fcf99a" path="/var/lib/kubelet/pods/ae6f1293-c2f8-4231-ae43-8c40d3fcf99a/volumes" Feb 15 21:38:19 crc kubenswrapper[4735]: I0215 21:38:19.680369 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:38:19 crc kubenswrapper[4735]: I0215 21:38:19.680776 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:38:49 crc kubenswrapper[4735]: I0215 21:38:49.680252 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:38:49 crc kubenswrapper[4735]: I0215 21:38:49.681654 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:39:05 crc kubenswrapper[4735]: I0215 21:39:05.273693 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d4fd7fbc4-nt694_c0df081d-fe76-415d-b26f-c67e2c97284f/barbican-api/0.log" Feb 15 21:39:05 crc kubenswrapper[4735]: I0215 21:39:05.423834 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d4fd7fbc4-nt694_c0df081d-fe76-415d-b26f-c67e2c97284f/barbican-api-log/0.log" Feb 15 21:39:06 crc kubenswrapper[4735]: I0215 21:39:06.120159 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f776895b6-v9kd8_8c4ffab0-9e1c-4ccb-841d-54c812217899/barbican-keystone-listener/0.log" Feb 15 21:39:06 crc kubenswrapper[4735]: I0215 21:39:06.148583 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f776895b6-v9kd8_8c4ffab0-9e1c-4ccb-841d-54c812217899/barbican-keystone-listener-log/0.log" Feb 15 21:39:06 crc kubenswrapper[4735]: I0215 21:39:06.193774 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b497d9d6c-4dqkk_29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2/barbican-worker/0.log" Feb 15 21:39:06 crc kubenswrapper[4735]: I0215 21:39:06.329085 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b497d9d6c-4dqkk_29c2633b-43d7-4a1d-9aed-1c29b0b2c6d2/barbican-worker-log/0.log" Feb 15 21:39:06 crc kubenswrapper[4735]: I0215 21:39:06.410426 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-95sqr_b5df6771-b0ed-45d4-aeac-219bb79caf7a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:06 crc kubenswrapper[4735]: I0215 21:39:06.612312 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_09208797-cc26-4216-b27f-51b281eba74a/ceilometer-central-agent/0.log" Feb 15 21:39:06 crc kubenswrapper[4735]: I0215 21:39:06.685394 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_09208797-cc26-4216-b27f-51b281eba74a/ceilometer-notification-agent/0.log" Feb 15 21:39:06 crc kubenswrapper[4735]: I0215 21:39:06.739154 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_09208797-cc26-4216-b27f-51b281eba74a/sg-core/0.log" Feb 15 21:39:06 crc kubenswrapper[4735]: I0215 21:39:06.826279 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_09208797-cc26-4216-b27f-51b281eba74a/proxy-httpd/0.log" Feb 15 21:39:07 crc kubenswrapper[4735]: I0215 21:39:07.318240 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1f3720a1-390e-457d-a4c2-d758b04a90d3/cinder-api/0.log" Feb 15 21:39:07 crc kubenswrapper[4735]: I0215 21:39:07.365162 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1f3720a1-390e-457d-a4c2-d758b04a90d3/cinder-api-log/0.log" Feb 15 21:39:07 crc kubenswrapper[4735]: I0215 21:39:07.435991 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7464049d-a3f2-44c0-982d-222196019aa2/cinder-scheduler/0.log" Feb 15 21:39:07 crc kubenswrapper[4735]: I0215 21:39:07.623880 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7464049d-a3f2-44c0-982d-222196019aa2/probe/0.log" Feb 15 21:39:07 crc kubenswrapper[4735]: I0215 21:39:07.635195 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-bwldn_f62f1800-36ef-497f-94af-ca96c2a3f7e9/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:07 crc kubenswrapper[4735]: I0215 21:39:07.892414 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-pnv2f_fd43082b-0db8-4620-9784-cc587548f737/init/0.log" Feb 15 21:39:07 crc kubenswrapper[4735]: I0215 21:39:07.898409 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5pc4x_4a254833-483c-4f57-8407-2dad01349aec/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:08 crc kubenswrapper[4735]: I0215 21:39:08.162152 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-pnv2f_fd43082b-0db8-4620-9784-cc587548f737/init/0.log" Feb 15 21:39:08 crc kubenswrapper[4735]: I0215 21:39:08.196510 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-jq8kg_a94a8020-82e8-4b37-add7-d9be8fc04c44/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:08 crc kubenswrapper[4735]: I0215 21:39:08.344659 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-pnv2f_fd43082b-0db8-4620-9784-cc587548f737/dnsmasq-dns/0.log" Feb 15 21:39:08 crc kubenswrapper[4735]: I0215 21:39:08.533093 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3a32ed00-6692-40df-af48-ecd34af2ffd2/glance-httpd/0.log" Feb 15 21:39:08 crc kubenswrapper[4735]: I0215 21:39:08.542473 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3a32ed00-6692-40df-af48-ecd34af2ffd2/glance-log/0.log" Feb 15 21:39:08 crc kubenswrapper[4735]: I0215 21:39:08.789655 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91998299-81cc-409a-a32d-2d2fa671d379/glance-httpd/0.log" Feb 15 21:39:08 crc kubenswrapper[4735]: I0215 21:39:08.802219 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91998299-81cc-409a-a32d-2d2fa671d379/glance-log/0.log" Feb 15 21:39:09 crc kubenswrapper[4735]: I0215 21:39:09.010031 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-c647fbf5b-7zs9r_970d5293-b621-45e1-a1e4-8cc176c9a148/horizon/2.log" Feb 15 21:39:09 crc kubenswrapper[4735]: I0215 21:39:09.198115 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-c647fbf5b-7zs9r_970d5293-b621-45e1-a1e4-8cc176c9a148/horizon/1.log" Feb 15 21:39:09 crc kubenswrapper[4735]: I0215 21:39:09.324608 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bxrmv_0ea9cfe8-7371-4d83-8e57-755d86809d46/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:09 crc kubenswrapper[4735]: I0215 21:39:09.470798 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-c647fbf5b-7zs9r_970d5293-b621-45e1-a1e4-8cc176c9a148/horizon-log/0.log" Feb 15 21:39:09 crc kubenswrapper[4735]: I0215 21:39:09.495626 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-c7dpc_38bf8db6-1cb0-49e1-b41f-ad45cee3d3a0/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:10 crc kubenswrapper[4735]: I0215 21:39:10.122476 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29519821-8vcgr_25286b3f-a5eb-49ce-a0dc-e2a8e99d9264/keystone-cron/0.log" Feb 15 21:39:10 crc kubenswrapper[4735]: I0215 21:39:10.344613 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a0faafd8-ba5f-4947-b5f5-e3b6e84a7f04/kube-state-metrics/0.log" Feb 15 21:39:10 crc kubenswrapper[4735]: I0215 21:39:10.372348 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7cb4db7f5b-mrtp9_33778124-7c75-4b6f-a2e1-c0ac0d8eda7e/keystone-api/0.log" Feb 15 21:39:10 crc kubenswrapper[4735]: I0215 21:39:10.459453 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ttx5g_3e63486b-442e-477f-8692-cb3da5e648cd/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:10 crc kubenswrapper[4735]: I0215 21:39:10.980341 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-rd8rs_82354908-424c-4069-abd7-f6b5ededdf13/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:11 crc kubenswrapper[4735]: I0215 21:39:11.100624 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-fdf549c7-xpft7_33cc02bd-6580-4d99-9c26-5f0c222e6461/neutron-api/0.log" Feb 15 21:39:11 crc kubenswrapper[4735]: I0215 21:39:11.138678 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-fdf549c7-xpft7_33cc02bd-6580-4d99-9c26-5f0c222e6461/neutron-httpd/0.log" Feb 15 21:39:11 crc kubenswrapper[4735]: I0215 21:39:11.924586 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d38fb2ae-a738-456f-b55e-65c70672dd2a/nova-cell0-conductor-conductor/0.log" Feb 15 21:39:12 crc kubenswrapper[4735]: I0215 21:39:12.177355 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_943d7959-9276-4940-8783-27492dfdf13e/nova-cell1-conductor-conductor/0.log" Feb 15 21:39:12 crc kubenswrapper[4735]: I0215 21:39:12.486818 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a964a23e-2012-4d43-98c2-e0b15f9d9d74/nova-api-log/0.log" Feb 15 21:39:12 crc kubenswrapper[4735]: I0215 21:39:12.651120 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_cc100b91-f66d-43d2-8b63-d52a3c27769f/nova-cell1-novncproxy-novncproxy/0.log" Feb 15 21:39:12 crc kubenswrapper[4735]: I0215 21:39:12.744540 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-t2s42_01424332-5cb6-4959-90b3-eb98af1ee96e/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:12 crc kubenswrapper[4735]: I0215 21:39:12.990548 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a964a23e-2012-4d43-98c2-e0b15f9d9d74/nova-api-api/0.log" Feb 15 21:39:13 crc kubenswrapper[4735]: I0215 21:39:13.034832 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_90c9c199-a27c-464a-8845-c2fd4e0b0fd9/nova-metadata-log/0.log" Feb 15 21:39:13 crc kubenswrapper[4735]: I0215 21:39:13.520983 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1f87d118-86d0-477e-a735-d2cd281b334a/mysql-bootstrap/0.log" Feb 15 21:39:13 crc kubenswrapper[4735]: I0215 21:39:13.857238 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d2066599-479d-43c3-86cb-528fc3922810/nova-scheduler-scheduler/0.log" Feb 15 21:39:14 crc kubenswrapper[4735]: I0215 21:39:14.067253 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1f87d118-86d0-477e-a735-d2cd281b334a/galera/0.log" Feb 15 21:39:14 crc kubenswrapper[4735]: I0215 21:39:14.081821 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1f87d118-86d0-477e-a735-d2cd281b334a/mysql-bootstrap/0.log" Feb 15 21:39:14 crc kubenswrapper[4735]: I0215 21:39:14.349236 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aa84f1ba-392a-4523-927c-405fab5bd619/mysql-bootstrap/0.log" Feb 15 21:39:14 crc kubenswrapper[4735]: I0215 21:39:14.549205 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aa84f1ba-392a-4523-927c-405fab5bd619/mysql-bootstrap/0.log" Feb 15 21:39:14 crc kubenswrapper[4735]: I0215 21:39:14.861910 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0e97b2d7-86cc-4ba1-9bb9-01b0dae7db15/openstackclient/0.log" Feb 15 21:39:15 crc kubenswrapper[4735]: I0215 21:39:15.125763 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aa84f1ba-392a-4523-927c-405fab5bd619/galera/0.log" Feb 15 21:39:15 crc kubenswrapper[4735]: I0215 21:39:15.576343 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_90c9c199-a27c-464a-8845-c2fd4e0b0fd9/nova-metadata-metadata/0.log" Feb 15 21:39:15 crc kubenswrapper[4735]: I0215 21:39:15.588112 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-8vw9j_3ff080f2-642e-4ff4-9554-5489daca64c4/openstack-network-exporter/0.log" Feb 15 21:39:15 crc kubenswrapper[4735]: I0215 21:39:15.621779 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fgcfc_9ca538ce-5609-4a07-887c-b9757cd7ad37/ovsdb-server-init/0.log" Feb 15 21:39:15 crc kubenswrapper[4735]: I0215 21:39:15.939736 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fgcfc_9ca538ce-5609-4a07-887c-b9757cd7ad37/ovsdb-server-init/0.log" Feb 15 21:39:15 crc kubenswrapper[4735]: I0215 21:39:15.940179 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fgcfc_9ca538ce-5609-4a07-887c-b9757cd7ad37/ovs-vswitchd/0.log" Feb 15 21:39:15 crc kubenswrapper[4735]: I0215 21:39:15.980605 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fgcfc_9ca538ce-5609-4a07-887c-b9757cd7ad37/ovsdb-server/0.log" Feb 15 21:39:16 crc kubenswrapper[4735]: I0215 21:39:16.228852 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tk256_df76a192-fa00-4146-8f3a-8f7fa6798fec/ovn-controller/0.log" Feb 15 21:39:16 crc kubenswrapper[4735]: I0215 21:39:16.328787 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-qc8mw_5779edb8-01f9-4f91-81fe-9af70711a9bc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:16 crc kubenswrapper[4735]: I0215 21:39:16.484268 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d/openstack-network-exporter/0.log" Feb 15 21:39:16 crc kubenswrapper[4735]: I0215 21:39:16.577332 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f87a4a7-6fc0-4b1a-b7c9-e5cc019fe81d/ovn-northd/0.log" Feb 15 21:39:16 crc kubenswrapper[4735]: I0215 21:39:16.752609 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_77c1f16a-92e0-4c7a-a92c-12328f59f0b8/openstack-network-exporter/0.log" Feb 15 21:39:16 crc kubenswrapper[4735]: I0215 21:39:16.822534 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_77c1f16a-92e0-4c7a-a92c-12328f59f0b8/ovsdbserver-nb/0.log" Feb 15 21:39:16 crc kubenswrapper[4735]: I0215 21:39:16.941971 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8bfd475e-d45a-4d6e-8150-90da2bf08b6a/openstack-network-exporter/0.log" Feb 15 21:39:17 crc kubenswrapper[4735]: I0215 21:39:17.032481 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8bfd475e-d45a-4d6e-8150-90da2bf08b6a/ovsdbserver-sb/0.log" Feb 15 21:39:17 crc kubenswrapper[4735]: I0215 21:39:17.424792 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6c5c4f5b9b-sg99c_da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f/placement-api/0.log" Feb 15 21:39:17 crc kubenswrapper[4735]: I0215 21:39:17.467715 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b1105ffb-a8a0-44fd-8679-171e016f43b1/setup-container/0.log" Feb 15 21:39:17 crc kubenswrapper[4735]: I0215 21:39:17.556267 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6c5c4f5b9b-sg99c_da92ea2f-81cd-49f8-9ba9-52f0cae6eb0f/placement-log/0.log" Feb 15 21:39:17 crc kubenswrapper[4735]: I0215 21:39:17.957189 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b1105ffb-a8a0-44fd-8679-171e016f43b1/setup-container/0.log" Feb 15 21:39:18 crc kubenswrapper[4735]: I0215 21:39:18.219772 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40/setup-container/0.log" Feb 15 21:39:18 crc kubenswrapper[4735]: I0215 21:39:18.307744 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b1105ffb-a8a0-44fd-8679-171e016f43b1/rabbitmq/0.log" Feb 15 21:39:18 crc kubenswrapper[4735]: I0215 21:39:18.537797 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40/setup-container/0.log" Feb 15 21:39:18 crc kubenswrapper[4735]: I0215 21:39:18.562417 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ad4f32bb-f01d-4adb-8a96-f8f9c0ef2a40/rabbitmq/0.log" Feb 15 21:39:18 crc kubenswrapper[4735]: I0215 21:39:18.603806 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dllrf_7bbcab5d-3968-4e71-b53f-0c78e86b21cb/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:18 crc kubenswrapper[4735]: I0215 21:39:18.845660 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-df8c9_aad75417-7c89-443b-88ed-aad98d4109b2/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:18 crc kubenswrapper[4735]: I0215 21:39:18.989827 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-d4fr2_e31b2a4c-8571-4a0e-868f-f86b757785ac/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.161245 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9wkf7_a7ab7547-fa81-411c-b9dd-fed0f819cd41/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.307686 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-frlr4_08f45e83-df1b-430c-a2af-92b9370912a5/ssh-known-hosts-edpm-deployment/0.log" Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.609889 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b7b86987-zpgc8_843b8077-b23a-42f9-bf70-79767ce35b4f/proxy-httpd/0.log" Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.613274 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b7b86987-zpgc8_843b8077-b23a-42f9-bf70-79767ce35b4f/proxy-server/0.log" Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.681677 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.681725 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.681763 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.682588 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"79803829e7f502800603cb50ee8fbc1de3db60bf3e331d273572d882ec3710ef"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.682640 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://79803829e7f502800603cb50ee8fbc1de3db60bf3e331d273572d882ec3710ef" gracePeriod=600 Feb 15 21:39:19 crc kubenswrapper[4735]: I0215 21:39:19.991540 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/account-reaper/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.002154 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/account-auditor/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.018199 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-zrmz5_35e018a0-8bf3-4725-a229-5a4196d4ae97/swift-ring-rebalance/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.223501 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/container-auditor/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.229962 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/account-replicator/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.285214 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/account-server/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.435064 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/container-replicator/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.506682 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/container-updater/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.507219 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="79803829e7f502800603cb50ee8fbc1de3db60bf3e331d273572d882ec3710ef" exitCode=0 Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.507270 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"79803829e7f502800603cb50ee8fbc1de3db60bf3e331d273572d882ec3710ef"} Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.507302 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b"} Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.507323 4735 scope.go:117] "RemoveContainer" containerID="45fadf187956f675712a9c2bb1ea9a408fa1f2bd93ccade25a8126789de5fe91" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.542028 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-auditor/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.569817 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/container-server/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.698485 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-expirer/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.770529 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-server/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.818045 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-replicator/0.log" Feb 15 21:39:20 crc kubenswrapper[4735]: I0215 21:39:20.924683 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/object-updater/0.log" Feb 15 21:39:21 crc kubenswrapper[4735]: I0215 21:39:21.013071 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/swift-recon-cron/0.log" Feb 15 21:39:21 crc kubenswrapper[4735]: I0215 21:39:21.041123 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a9642a7f-e8aa-4368-b2b9-3da6537c4d8f/rsync/0.log" Feb 15 21:39:21 crc kubenswrapper[4735]: I0215 21:39:21.972829 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_abcf2b9d-8410-4a1b-be0e-3dde2a6e3383/tempest-tests-tempest-tests-runner/0.log" Feb 15 21:39:21 crc kubenswrapper[4735]: I0215 21:39:21.986784 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-fd8hf_25a254b5-abc4-45fa-aba5-935d79ae8959/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:22 crc kubenswrapper[4735]: I0215 21:39:22.256282 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_97a6ddef-7ccf-4c1f-bc90-691f4511e4ad/test-operator-logs-container/0.log" Feb 15 21:39:22 crc kubenswrapper[4735]: I0215 21:39:22.529082 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-lx2ch_6c7b61e8-5a6a-4651-bf3b-3f90ae858b40/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 15 21:39:34 crc kubenswrapper[4735]: I0215 21:39:34.569747 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6a34b649-4566-4a07-846e-e97886eec195/memcached/0.log" Feb 15 21:39:56 crc kubenswrapper[4735]: I0215 21:39:56.197929 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/util/0.log" Feb 15 21:39:56 crc kubenswrapper[4735]: I0215 21:39:56.427446 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/pull/0.log" Feb 15 21:39:56 crc kubenswrapper[4735]: I0215 21:39:56.467585 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/util/0.log" Feb 15 21:39:56 crc kubenswrapper[4735]: I0215 21:39:56.505323 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/pull/0.log" Feb 15 21:39:56 crc kubenswrapper[4735]: I0215 21:39:56.876828 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/util/0.log" Feb 15 21:39:56 crc kubenswrapper[4735]: I0215 21:39:56.922880 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/pull/0.log" Feb 15 21:39:56 crc kubenswrapper[4735]: I0215 21:39:56.941192 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8aad70f85ca0f9c867184019e01f168d491ff060fcd5e0bebb6495bc9emfldq_8ceaf5ed-e64d-4d6c-8990-2184dddaf8af/extract/0.log" Feb 15 21:39:57 crc kubenswrapper[4735]: I0215 21:39:57.362687 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-vddvx_99564ffe-1d8b-44b1-a625-4445d624a98f/manager/0.log" Feb 15 21:39:57 crc kubenswrapper[4735]: I0215 21:39:57.726817 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-gvnkh_92aeecb8-b9fa-4db8-916e-f2dd800d2da3/manager/0.log" Feb 15 21:39:58 crc kubenswrapper[4735]: I0215 21:39:58.003635 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-v4wpf_ab28d8d3-58f1-460c-a0ed-75f61ebc2c52/manager/0.log" Feb 15 21:39:58 crc kubenswrapper[4735]: I0215 21:39:58.464012 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-fslrr_31865132-118d-4f4c-bf0b-44ab13060882/manager/0.log" Feb 15 21:39:59 crc kubenswrapper[4735]: I0215 21:39:59.230296 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7676946c7c-5w649_79556e2d-fb3e-4a41-a85c-38005174ba50/manager/0.log" Feb 15 21:39:59 crc kubenswrapper[4735]: I0215 21:39:59.247428 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-rbhc6_2b6ae856-4502-44cd-ad23-21577792b6a0/manager/0.log" Feb 15 21:39:59 crc kubenswrapper[4735]: I0215 21:39:59.478832 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-v6nvk_98e57a69-0860-468e-898c-9ec42836d3d4/manager/0.log" Feb 15 21:39:59 crc kubenswrapper[4735]: I0215 21:39:59.590584 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-ghs7m_8802a6f8-206d-473d-854c-77e3c74389b5/manager/0.log" Feb 15 21:39:59 crc kubenswrapper[4735]: I0215 21:39:59.848179 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-mh89r_1f526ee2-4ba8-424c-b3f0-dbcc357c6ea0/manager/0.log" Feb 15 21:40:00 crc kubenswrapper[4735]: I0215 21:40:00.164709 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-4hlsd_1094ce7c-b104-40d4-8e0f-3ee695bfd6cd/manager/0.log" Feb 15 21:40:00 crc kubenswrapper[4735]: I0215 21:40:00.210635 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-jwrfj_79e82794-8193-4eba-af98-f6e0ae43274b/manager/0.log" Feb 15 21:40:00 crc kubenswrapper[4735]: I0215 21:40:00.567814 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-t9nc4_d8c734e9-fd8c-459d-982f-53b0a706d0a0/manager/0.log" Feb 15 21:40:00 crc kubenswrapper[4735]: I0215 21:40:00.783917 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84966cf5c48cdg5_9549fdaf-fe4c-4da2-92db-db91beb95450/manager/0.log" Feb 15 21:40:01 crc kubenswrapper[4735]: I0215 21:40:01.292423 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-58498c75bf-2kzsp_e8c2edc3-d324-4b8e-9831-fcf924f6a655/operator/0.log" Feb 15 21:40:01 crc kubenswrapper[4735]: I0215 21:40:01.490201 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-ldbt6_32c1d680-92cf-4c49-93b9-17bb17aaf623/registry-server/0.log" Feb 15 21:40:01 crc kubenswrapper[4735]: I0215 21:40:01.809589 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-f8s72_1bd0ae4a-c64a-4f7c-a7c5-4e9b683a9aea/manager/0.log" Feb 15 21:40:02 crc kubenswrapper[4735]: I0215 21:40:02.079564 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-9trcb_d8bf1d9d-df22-4b17-91b7-833052b5a337/manager/0.log" Feb 15 21:40:02 crc kubenswrapper[4735]: I0215 21:40:02.292744 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-9z8vz_e57b4eea-5c8f-4caa-9644-0994108974da/operator/0.log" Feb 15 21:40:02 crc kubenswrapper[4735]: I0215 21:40:02.557164 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-vbr6n_14de9f25-fdbd-4d44-9f2c-440b256c4faf/manager/0.log" Feb 15 21:40:02 crc kubenswrapper[4735]: I0215 21:40:02.884405 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-lfj2m_4fb2a283-35ca-4391-8e59-79a8fa7bb4f0/manager/0.log" Feb 15 21:40:03 crc kubenswrapper[4735]: I0215 21:40:03.118355 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-r6bz7_3fa6a777-e009-4e48-b05d-ac9b34acd21c/manager/0.log" Feb 15 21:40:03 crc kubenswrapper[4735]: I0215 21:40:03.332241 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79f9b74c8-4b426_6fad7194-6e10-4dc9-ac3c-241722357e55/manager/0.log" Feb 15 21:40:03 crc kubenswrapper[4735]: I0215 21:40:03.349355 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-d65l6_26e7f3b0-6f34-491e-917a-e545d320ee89/manager/0.log" Feb 15 21:40:03 crc kubenswrapper[4735]: I0215 21:40:03.757408 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-w8vz9_c7d8cf5f-13c8-4857-96d2-0da2c88ed221/manager/0.log" Feb 15 21:40:08 crc kubenswrapper[4735]: I0215 21:40:08.157795 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-9hc4f_b9be4a5f-fbec-4e52-bb00-cbc08576e1fd/manager/0.log" Feb 15 21:40:28 crc kubenswrapper[4735]: I0215 21:40:28.728196 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-4wxcd_3036a2ba-e736-416a-bcf1-684d246cd3f1/control-plane-machine-set-operator/0.log" Feb 15 21:40:29 crc kubenswrapper[4735]: I0215 21:40:29.005757 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rjhgs_34482250-9472-45fe-aab5-485c7f129bca/kube-rbac-proxy/0.log" Feb 15 21:40:29 crc kubenswrapper[4735]: I0215 21:40:29.007723 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rjhgs_34482250-9472-45fe-aab5-485c7f129bca/machine-api-operator/0.log" Feb 15 21:40:44 crc kubenswrapper[4735]: I0215 21:40:44.814604 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-2jh8n_56f25e9b-225b-433d-af08-bdd6bb37e6e4/cert-manager-controller/0.log" Feb 15 21:40:45 crc kubenswrapper[4735]: I0215 21:40:45.019187 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-88v64_e04e2b12-a2ee-4ba4-acb4-0276c777a066/cert-manager-cainjector/0.log" Feb 15 21:40:45 crc kubenswrapper[4735]: I0215 21:40:45.133669 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-jvntw_468178fa-6b3c-41b4-9bf2-45178ebfb3c2/cert-manager-webhook/0.log" Feb 15 21:41:00 crc kubenswrapper[4735]: I0215 21:41:00.373084 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-6q5t8_4d5d6279-ce2c-4382-900c-77563978d63a/nmstate-console-plugin/0.log" Feb 15 21:41:00 crc kubenswrapper[4735]: I0215 21:41:00.641344 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5grq7_ffa58b6c-ce4f-494b-99b7-a0d45b2b8b8e/nmstate-handler/0.log" Feb 15 21:41:00 crc kubenswrapper[4735]: I0215 21:41:00.725175 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-4gnfl_6291dd59-7c75-4426-86d1-063a72855309/kube-rbac-proxy/0.log" Feb 15 21:41:00 crc kubenswrapper[4735]: I0215 21:41:00.794567 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-4gnfl_6291dd59-7c75-4426-86d1-063a72855309/nmstate-metrics/0.log" Feb 15 21:41:00 crc kubenswrapper[4735]: I0215 21:41:00.953429 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-vc7vf_16c72a4d-04eb-4a5f-af25-439d36f5c9de/nmstate-operator/0.log" Feb 15 21:41:01 crc kubenswrapper[4735]: I0215 21:41:01.060673 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-7gkdk_cecb1f2e-e07b-4ea1-924d-7802fa40fd6b/nmstate-webhook/0.log" Feb 15 21:41:29 crc kubenswrapper[4735]: I0215 21:41:29.339744 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-7xh6x_ca5b18f3-83a4-497d-913b-e0539a6991b5/kube-rbac-proxy/0.log" Feb 15 21:41:29 crc kubenswrapper[4735]: I0215 21:41:29.349916 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-7xh6x_ca5b18f3-83a4-497d-913b-e0539a6991b5/controller/0.log" Feb 15 21:41:29 crc kubenswrapper[4735]: I0215 21:41:29.541386 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-frr-files/0.log" Feb 15 21:41:29 crc kubenswrapper[4735]: I0215 21:41:29.735804 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-frr-files/0.log" Feb 15 21:41:29 crc kubenswrapper[4735]: I0215 21:41:29.746083 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-reloader/0.log" Feb 15 21:41:29 crc kubenswrapper[4735]: I0215 21:41:29.786569 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-metrics/0.log" Feb 15 21:41:29 crc kubenswrapper[4735]: I0215 21:41:29.833053 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-reloader/0.log" Feb 15 21:41:29 crc kubenswrapper[4735]: I0215 21:41:29.990044 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-frr-files/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.018192 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-reloader/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.071387 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-metrics/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.084249 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-metrics/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.233036 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-reloader/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.270070 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/controller/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.272873 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-frr-files/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.278488 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/cp-metrics/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.431586 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/frr-metrics/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.469754 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/kube-rbac-proxy-frr/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.483736 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/kube-rbac-proxy/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.667767 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/reloader/0.log" Feb 15 21:41:30 crc kubenswrapper[4735]: I0215 21:41:30.750378 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-7km29_99caf845-7f1f-4a29-a8e5-c939ffb15c9e/frr-k8s-webhook-server/0.log" Feb 15 21:41:31 crc kubenswrapper[4735]: I0215 21:41:31.058990 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79df65bf49-hzgd7_c27da055-adee-4524-93ce-e2e83cf5966a/manager/0.log" Feb 15 21:41:31 crc kubenswrapper[4735]: I0215 21:41:31.188594 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7f47cf9495-7nwhm_3516fdcd-aa1a-48ea-9f44-b928ca524e28/webhook-server/0.log" Feb 15 21:41:31 crc kubenswrapper[4735]: I0215 21:41:31.381771 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9tpjm_ec4ea5be-56b7-42e2-b462-a880d12145af/kube-rbac-proxy/0.log" Feb 15 21:41:31 crc kubenswrapper[4735]: I0215 21:41:31.786170 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g5jlv_d5d3c0c9-8711-4346-9591-24807ee906ef/frr/0.log" Feb 15 21:41:31 crc kubenswrapper[4735]: I0215 21:41:31.874764 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9tpjm_ec4ea5be-56b7-42e2-b462-a880d12145af/speaker/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.005075 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/util/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.243581 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/util/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.244223 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/pull/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.298492 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/pull/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.442558 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/extract/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.445926 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/util/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.487134 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc21375ckc_9afae777-8644-4e09-8166-55b3e1e34de9/pull/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.621669 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-utilities/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.801237 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-utilities/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.811053 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-content/0.log" Feb 15 21:41:46 crc kubenswrapper[4735]: I0215 21:41:46.852288 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-content/0.log" Feb 15 21:41:47 crc kubenswrapper[4735]: I0215 21:41:47.026163 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-content/0.log" Feb 15 21:41:47 crc kubenswrapper[4735]: I0215 21:41:47.031822 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/extract-utilities/0.log" Feb 15 21:41:47 crc kubenswrapper[4735]: I0215 21:41:47.273925 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-utilities/0.log" Feb 15 21:41:47 crc kubenswrapper[4735]: I0215 21:41:47.533509 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d4x4r_af33352e-df7a-4fb2-af44-d084a141e402/registry-server/0.log" Feb 15 21:41:47 crc kubenswrapper[4735]: I0215 21:41:47.606124 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-content/0.log" Feb 15 21:41:47 crc kubenswrapper[4735]: I0215 21:41:47.606871 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-utilities/0.log" Feb 15 21:41:47 crc kubenswrapper[4735]: I0215 21:41:47.617069 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-content/0.log" Feb 15 21:41:47 crc kubenswrapper[4735]: I0215 21:41:47.765835 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-utilities/0.log" Feb 15 21:41:47 crc kubenswrapper[4735]: I0215 21:41:47.815514 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/extract-content/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.063218 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/util/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.297984 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/util/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.326832 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkthb_ceaf5522-aed5-418f-b483-4c0b945dc615/registry-server/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.377101 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/pull/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.395531 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/pull/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.589393 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/pull/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.612307 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/util/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.642106 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecafqgln_880dc366-4d49-4093-b67d-715fea45d07a/extract/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.870498 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ff4s6_02a0669b-e47e-4f62-851c-a776d1f43b12/marketplace-operator/0.log" Feb 15 21:41:48 crc kubenswrapper[4735]: I0215 21:41:48.877914 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-utilities/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.128333 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-utilities/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.144543 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-content/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.164682 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-content/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.287412 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-utilities/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.341565 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/extract-content/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.498496 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gvxkv_f41f0acf-7b74-44f7-92a4-30fb79ccea7a/registry-server/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.614516 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-utilities/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.680022 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.680086 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.719717 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-utilities/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.722646 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-content/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.771258 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-content/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.976811 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-content/0.log" Feb 15 21:41:49 crc kubenswrapper[4735]: I0215 21:41:49.991627 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/extract-utilities/0.log" Feb 15 21:41:50 crc kubenswrapper[4735]: I0215 21:41:50.553397 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4stk_aec1828e-1499-4e18-8b2d-5cfbe9a6d457/registry-server/0.log" Feb 15 21:42:19 crc kubenswrapper[4735]: I0215 21:42:19.679922 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:42:19 crc kubenswrapper[4735]: I0215 21:42:19.680433 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:42:49 crc kubenswrapper[4735]: I0215 21:42:49.680704 4735 patch_prober.go:28] interesting pod/machine-config-daemon-kjps6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 15 21:42:49 crc kubenswrapper[4735]: I0215 21:42:49.681379 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 15 21:42:49 crc kubenswrapper[4735]: I0215 21:42:49.681434 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" Feb 15 21:42:49 crc kubenswrapper[4735]: I0215 21:42:49.683702 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b"} pod="openshift-machine-config-operator/machine-config-daemon-kjps6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 15 21:42:49 crc kubenswrapper[4735]: I0215 21:42:49.683795 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerName="machine-config-daemon" containerID="cri-o://b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" gracePeriod=600 Feb 15 21:42:49 crc kubenswrapper[4735]: E0215 21:42:49.810460 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:42:50 crc kubenswrapper[4735]: I0215 21:42:50.366861 4735 generic.go:334] "Generic (PLEG): container finished" podID="43de4639-a922-4182-9377-d1f28b3aa4c1" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" exitCode=0 Feb 15 21:42:50 crc kubenswrapper[4735]: I0215 21:42:50.366934 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerDied","Data":"b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b"} Feb 15 21:42:50 crc kubenswrapper[4735]: I0215 21:42:50.367357 4735 scope.go:117] "RemoveContainer" containerID="79803829e7f502800603cb50ee8fbc1de3db60bf3e331d273572d882ec3710ef" Feb 15 21:42:50 crc kubenswrapper[4735]: I0215 21:42:50.369319 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:42:50 crc kubenswrapper[4735]: E0215 21:42:50.370601 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:43:03 crc kubenswrapper[4735]: I0215 21:43:03.887398 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:43:03 crc kubenswrapper[4735]: E0215 21:43:03.888393 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:43:15 crc kubenswrapper[4735]: I0215 21:43:15.887879 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:43:15 crc kubenswrapper[4735]: E0215 21:43:15.888975 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.256419 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g7zbx"] Feb 15 21:43:23 crc kubenswrapper[4735]: E0215 21:43:23.257524 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae6f1293-c2f8-4231-ae43-8c40d3fcf99a" containerName="container-00" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.257540 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae6f1293-c2f8-4231-ae43-8c40d3fcf99a" containerName="container-00" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.257774 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae6f1293-c2f8-4231-ae43-8c40d3fcf99a" containerName="container-00" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.260987 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.270894 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g7zbx"] Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.370044 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tzmb\" (UniqueName: \"kubernetes.io/projected/59593045-41dc-4bfa-aafe-b2044c8b6797-kube-api-access-7tzmb\") pod \"certified-operators-g7zbx\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.370167 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-catalog-content\") pod \"certified-operators-g7zbx\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.370331 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-utilities\") pod \"certified-operators-g7zbx\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.472819 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tzmb\" (UniqueName: \"kubernetes.io/projected/59593045-41dc-4bfa-aafe-b2044c8b6797-kube-api-access-7tzmb\") pod \"certified-operators-g7zbx\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.473381 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-catalog-content\") pod \"certified-operators-g7zbx\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.473617 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-utilities\") pod \"certified-operators-g7zbx\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.473820 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-catalog-content\") pod \"certified-operators-g7zbx\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.474098 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-utilities\") pod \"certified-operators-g7zbx\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.491286 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tzmb\" (UniqueName: \"kubernetes.io/projected/59593045-41dc-4bfa-aafe-b2044c8b6797-kube-api-access-7tzmb\") pod \"certified-operators-g7zbx\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:23 crc kubenswrapper[4735]: I0215 21:43:23.598914 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:24 crc kubenswrapper[4735]: I0215 21:43:24.146222 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g7zbx"] Feb 15 21:43:25 crc kubenswrapper[4735]: I0215 21:43:25.153808 4735 generic.go:334] "Generic (PLEG): container finished" podID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerID="ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c" exitCode=0 Feb 15 21:43:25 crc kubenswrapper[4735]: I0215 21:43:25.153921 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7zbx" event={"ID":"59593045-41dc-4bfa-aafe-b2044c8b6797","Type":"ContainerDied","Data":"ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c"} Feb 15 21:43:25 crc kubenswrapper[4735]: I0215 21:43:25.154158 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7zbx" event={"ID":"59593045-41dc-4bfa-aafe-b2044c8b6797","Type":"ContainerStarted","Data":"e28a55f937354dbd7f58745e18b3296f0d31e54f61572c747c7df96c63b888d6"} Feb 15 21:43:25 crc kubenswrapper[4735]: I0215 21:43:25.158197 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 15 21:43:26 crc kubenswrapper[4735]: I0215 21:43:26.168166 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7zbx" event={"ID":"59593045-41dc-4bfa-aafe-b2044c8b6797","Type":"ContainerStarted","Data":"1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480"} Feb 15 21:43:27 crc kubenswrapper[4735]: I0215 21:43:27.179807 4735 generic.go:334] "Generic (PLEG): container finished" podID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerID="1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480" exitCode=0 Feb 15 21:43:27 crc kubenswrapper[4735]: I0215 21:43:27.179854 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7zbx" event={"ID":"59593045-41dc-4bfa-aafe-b2044c8b6797","Type":"ContainerDied","Data":"1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480"} Feb 15 21:43:28 crc kubenswrapper[4735]: I0215 21:43:28.199266 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7zbx" event={"ID":"59593045-41dc-4bfa-aafe-b2044c8b6797","Type":"ContainerStarted","Data":"e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de"} Feb 15 21:43:28 crc kubenswrapper[4735]: I0215 21:43:28.229134 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g7zbx" podStartSLOduration=2.776328449 podStartE2EDuration="5.229117313s" podCreationTimestamp="2026-02-15 21:43:23 +0000 UTC" firstStartedPulling="2026-02-15 21:43:25.157977686 +0000 UTC m=+5213.023993309" lastFinishedPulling="2026-02-15 21:43:27.61076654 +0000 UTC m=+5215.476782173" observedRunningTime="2026-02-15 21:43:28.226057861 +0000 UTC m=+5216.092073484" watchObservedRunningTime="2026-02-15 21:43:28.229117313 +0000 UTC m=+5216.095132946" Feb 15 21:43:30 crc kubenswrapper[4735]: I0215 21:43:30.890890 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:43:30 crc kubenswrapper[4735]: E0215 21:43:30.891618 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:43:33 crc kubenswrapper[4735]: I0215 21:43:33.600568 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:33 crc kubenswrapper[4735]: I0215 21:43:33.603142 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:33 crc kubenswrapper[4735]: I0215 21:43:33.668124 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:34 crc kubenswrapper[4735]: I0215 21:43:34.326482 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:34 crc kubenswrapper[4735]: I0215 21:43:34.378161 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g7zbx"] Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.285194 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g7zbx" podUID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerName="registry-server" containerID="cri-o://e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de" gracePeriod=2 Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.820219 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.851639 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tzmb\" (UniqueName: \"kubernetes.io/projected/59593045-41dc-4bfa-aafe-b2044c8b6797-kube-api-access-7tzmb\") pod \"59593045-41dc-4bfa-aafe-b2044c8b6797\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.851818 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-utilities\") pod \"59593045-41dc-4bfa-aafe-b2044c8b6797\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.851894 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-catalog-content\") pod \"59593045-41dc-4bfa-aafe-b2044c8b6797\" (UID: \"59593045-41dc-4bfa-aafe-b2044c8b6797\") " Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.874561 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-utilities" (OuterVolumeSpecName: "utilities") pod "59593045-41dc-4bfa-aafe-b2044c8b6797" (UID: "59593045-41dc-4bfa-aafe-b2044c8b6797"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.881717 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59593045-41dc-4bfa-aafe-b2044c8b6797-kube-api-access-7tzmb" (OuterVolumeSpecName: "kube-api-access-7tzmb") pod "59593045-41dc-4bfa-aafe-b2044c8b6797" (UID: "59593045-41dc-4bfa-aafe-b2044c8b6797"). InnerVolumeSpecName "kube-api-access-7tzmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.927197 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59593045-41dc-4bfa-aafe-b2044c8b6797" (UID: "59593045-41dc-4bfa-aafe-b2044c8b6797"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.967479 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.967505 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59593045-41dc-4bfa-aafe-b2044c8b6797-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:43:36 crc kubenswrapper[4735]: I0215 21:43:36.967516 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tzmb\" (UniqueName: \"kubernetes.io/projected/59593045-41dc-4bfa-aafe-b2044c8b6797-kube-api-access-7tzmb\") on node \"crc\" DevicePath \"\"" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.296523 4735 generic.go:334] "Generic (PLEG): container finished" podID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerID="e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de" exitCode=0 Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.296566 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7zbx" event={"ID":"59593045-41dc-4bfa-aafe-b2044c8b6797","Type":"ContainerDied","Data":"e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de"} Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.296593 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7zbx" event={"ID":"59593045-41dc-4bfa-aafe-b2044c8b6797","Type":"ContainerDied","Data":"e28a55f937354dbd7f58745e18b3296f0d31e54f61572c747c7df96c63b888d6"} Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.296610 4735 scope.go:117] "RemoveContainer" containerID="e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.296630 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g7zbx" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.315379 4735 scope.go:117] "RemoveContainer" containerID="1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.335663 4735 scope.go:117] "RemoveContainer" containerID="ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.400472 4735 scope.go:117] "RemoveContainer" containerID="e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de" Feb 15 21:43:37 crc kubenswrapper[4735]: E0215 21:43:37.400938 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de\": container with ID starting with e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de not found: ID does not exist" containerID="e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.400993 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de"} err="failed to get container status \"e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de\": rpc error: code = NotFound desc = could not find container \"e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de\": container with ID starting with e50bc6c5e6bdfcd02ae2b1a9148b8c4a84804544e2251c0b856147cfca07d2de not found: ID does not exist" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.401017 4735 scope.go:117] "RemoveContainer" containerID="1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480" Feb 15 21:43:37 crc kubenswrapper[4735]: E0215 21:43:37.401351 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480\": container with ID starting with 1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480 not found: ID does not exist" containerID="1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.401381 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480"} err="failed to get container status \"1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480\": rpc error: code = NotFound desc = could not find container \"1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480\": container with ID starting with 1229414a185861882f5b496be35ed90dc55cba0301fd416f2345102c6a90e480 not found: ID does not exist" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.401402 4735 scope.go:117] "RemoveContainer" containerID="ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c" Feb 15 21:43:37 crc kubenswrapper[4735]: E0215 21:43:37.401734 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c\": container with ID starting with ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c not found: ID does not exist" containerID="ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.401758 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c"} err="failed to get container status \"ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c\": rpc error: code = NotFound desc = could not find container \"ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c\": container with ID starting with ce1ec224aaf66bff9712b90c3bbcb9b4e877c75c40c0afb576f824758e29287c not found: ID does not exist" Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.407098 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g7zbx"] Feb 15 21:43:37 crc kubenswrapper[4735]: I0215 21:43:37.422916 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g7zbx"] Feb 15 21:43:38 crc kubenswrapper[4735]: I0215 21:43:38.906412 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59593045-41dc-4bfa-aafe-b2044c8b6797" path="/var/lib/kubelet/pods/59593045-41dc-4bfa-aafe-b2044c8b6797/volumes" Feb 15 21:43:44 crc kubenswrapper[4735]: I0215 21:43:44.887977 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:43:44 crc kubenswrapper[4735]: E0215 21:43:44.888779 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:43:54 crc kubenswrapper[4735]: I0215 21:43:54.018455 4735 scope.go:117] "RemoveContainer" containerID="768e0f78a548966702f5215b4961fb1e735a5cffea5d303c4887b6ae0ae4107b" Feb 15 21:43:59 crc kubenswrapper[4735]: I0215 21:43:59.887277 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:43:59 crc kubenswrapper[4735]: E0215 21:43:59.887967 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:44:11 crc kubenswrapper[4735]: I0215 21:44:11.887403 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:44:11 crc kubenswrapper[4735]: E0215 21:44:11.888138 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:44:16 crc kubenswrapper[4735]: I0215 21:44:16.762494 4735 generic.go:334] "Generic (PLEG): container finished" podID="f92d063b-e0fa-401e-9a93-30fadc3e2425" containerID="bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b" exitCode=0 Feb 15 21:44:16 crc kubenswrapper[4735]: I0215 21:44:16.762580 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvx7p/must-gather-7657q" event={"ID":"f92d063b-e0fa-401e-9a93-30fadc3e2425","Type":"ContainerDied","Data":"bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b"} Feb 15 21:44:16 crc kubenswrapper[4735]: I0215 21:44:16.764123 4735 scope.go:117] "RemoveContainer" containerID="bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b" Feb 15 21:44:17 crc kubenswrapper[4735]: I0215 21:44:17.127910 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tvx7p_must-gather-7657q_f92d063b-e0fa-401e-9a93-30fadc3e2425/gather/0.log" Feb 15 21:44:22 crc kubenswrapper[4735]: I0215 21:44:22.892517 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:44:22 crc kubenswrapper[4735]: E0215 21:44:22.893633 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.122623 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tvx7p/must-gather-7657q"] Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.124755 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-tvx7p/must-gather-7657q" podUID="f92d063b-e0fa-401e-9a93-30fadc3e2425" containerName="copy" containerID="cri-o://81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d" gracePeriod=2 Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.135826 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tvx7p/must-gather-7657q"] Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.527791 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tvx7p_must-gather-7657q_f92d063b-e0fa-401e-9a93-30fadc3e2425/copy/0.log" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.528463 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.653175 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2hn7\" (UniqueName: \"kubernetes.io/projected/f92d063b-e0fa-401e-9a93-30fadc3e2425-kube-api-access-c2hn7\") pod \"f92d063b-e0fa-401e-9a93-30fadc3e2425\" (UID: \"f92d063b-e0fa-401e-9a93-30fadc3e2425\") " Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.654356 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f92d063b-e0fa-401e-9a93-30fadc3e2425-must-gather-output\") pod \"f92d063b-e0fa-401e-9a93-30fadc3e2425\" (UID: \"f92d063b-e0fa-401e-9a93-30fadc3e2425\") " Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.665662 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f92d063b-e0fa-401e-9a93-30fadc3e2425-kube-api-access-c2hn7" (OuterVolumeSpecName: "kube-api-access-c2hn7") pod "f92d063b-e0fa-401e-9a93-30fadc3e2425" (UID: "f92d063b-e0fa-401e-9a93-30fadc3e2425"). InnerVolumeSpecName "kube-api-access-c2hn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.756640 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2hn7\" (UniqueName: \"kubernetes.io/projected/f92d063b-e0fa-401e-9a93-30fadc3e2425-kube-api-access-c2hn7\") on node \"crc\" DevicePath \"\"" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.881132 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f92d063b-e0fa-401e-9a93-30fadc3e2425-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "f92d063b-e0fa-401e-9a93-30fadc3e2425" (UID: "f92d063b-e0fa-401e-9a93-30fadc3e2425"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.900247 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f92d063b-e0fa-401e-9a93-30fadc3e2425" path="/var/lib/kubelet/pods/f92d063b-e0fa-401e-9a93-30fadc3e2425/volumes" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.911969 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tvx7p_must-gather-7657q_f92d063b-e0fa-401e-9a93-30fadc3e2425/copy/0.log" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.913179 4735 generic.go:334] "Generic (PLEG): container finished" podID="f92d063b-e0fa-401e-9a93-30fadc3e2425" containerID="81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d" exitCode=143 Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.913231 4735 scope.go:117] "RemoveContainer" containerID="81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.913349 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvx7p/must-gather-7657q" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.962565 4735 scope.go:117] "RemoveContainer" containerID="bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b" Feb 15 21:44:30 crc kubenswrapper[4735]: I0215 21:44:30.964580 4735 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f92d063b-e0fa-401e-9a93-30fadc3e2425-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 15 21:44:31 crc kubenswrapper[4735]: I0215 21:44:31.020326 4735 scope.go:117] "RemoveContainer" containerID="81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d" Feb 15 21:44:31 crc kubenswrapper[4735]: E0215 21:44:31.020757 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d\": container with ID starting with 81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d not found: ID does not exist" containerID="81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d" Feb 15 21:44:31 crc kubenswrapper[4735]: I0215 21:44:31.020805 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d"} err="failed to get container status \"81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d\": rpc error: code = NotFound desc = could not find container \"81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d\": container with ID starting with 81efc7460f75c20fc13548cd68a901dbd3ec67cd40fe60dc528c30ab1c72bb3d not found: ID does not exist" Feb 15 21:44:31 crc kubenswrapper[4735]: I0215 21:44:31.020830 4735 scope.go:117] "RemoveContainer" containerID="bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b" Feb 15 21:44:31 crc kubenswrapper[4735]: E0215 21:44:31.021098 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b\": container with ID starting with bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b not found: ID does not exist" containerID="bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b" Feb 15 21:44:31 crc kubenswrapper[4735]: I0215 21:44:31.021133 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b"} err="failed to get container status \"bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b\": rpc error: code = NotFound desc = could not find container \"bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b\": container with ID starting with bc5921cefff837b8c1c6bfb0341ee389c93d64a175f07e3204b40bb09c5e0d5b not found: ID does not exist" Feb 15 21:44:33 crc kubenswrapper[4735]: I0215 21:44:33.886709 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:44:33 crc kubenswrapper[4735]: E0215 21:44:33.887247 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:44:44 crc kubenswrapper[4735]: I0215 21:44:44.888049 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:44:44 crc kubenswrapper[4735]: E0215 21:44:44.889107 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:44:54 crc kubenswrapper[4735]: I0215 21:44:54.119269 4735 scope.go:117] "RemoveContainer" containerID="07e11bb215e64d4a85fb253da1d5e41cd3f16b9bbdefe9889bbad9160fa7dbd0" Feb 15 21:44:55 crc kubenswrapper[4735]: I0215 21:44:55.886852 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:44:55 crc kubenswrapper[4735]: E0215 21:44:55.887440 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.158279 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z"] Feb 15 21:45:00 crc kubenswrapper[4735]: E0215 21:45:00.159539 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerName="registry-server" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.159564 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerName="registry-server" Feb 15 21:45:00 crc kubenswrapper[4735]: E0215 21:45:00.159595 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92d063b-e0fa-401e-9a93-30fadc3e2425" containerName="gather" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.159606 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92d063b-e0fa-401e-9a93-30fadc3e2425" containerName="gather" Feb 15 21:45:00 crc kubenswrapper[4735]: E0215 21:45:00.159632 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerName="extract-content" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.159645 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerName="extract-content" Feb 15 21:45:00 crc kubenswrapper[4735]: E0215 21:45:00.159676 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerName="extract-utilities" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.159687 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerName="extract-utilities" Feb 15 21:45:00 crc kubenswrapper[4735]: E0215 21:45:00.159707 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92d063b-e0fa-401e-9a93-30fadc3e2425" containerName="copy" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.159718 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92d063b-e0fa-401e-9a93-30fadc3e2425" containerName="copy" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.160064 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="59593045-41dc-4bfa-aafe-b2044c8b6797" containerName="registry-server" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.160113 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f92d063b-e0fa-401e-9a93-30fadc3e2425" containerName="copy" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.160132 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f92d063b-e0fa-401e-9a93-30fadc3e2425" containerName="gather" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.161163 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.166576 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.166585 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.184612 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z"] Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.226467 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fed7859-0158-469a-bec9-275064e23cf6-secret-volume\") pod \"collect-profiles-29519865-w9n5z\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.226559 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fed7859-0158-469a-bec9-275064e23cf6-config-volume\") pod \"collect-profiles-29519865-w9n5z\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.226582 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkr4n\" (UniqueName: \"kubernetes.io/projected/5fed7859-0158-469a-bec9-275064e23cf6-kube-api-access-dkr4n\") pod \"collect-profiles-29519865-w9n5z\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.327871 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fed7859-0158-469a-bec9-275064e23cf6-secret-volume\") pod \"collect-profiles-29519865-w9n5z\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.327964 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fed7859-0158-469a-bec9-275064e23cf6-config-volume\") pod \"collect-profiles-29519865-w9n5z\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.327989 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkr4n\" (UniqueName: \"kubernetes.io/projected/5fed7859-0158-469a-bec9-275064e23cf6-kube-api-access-dkr4n\") pod \"collect-profiles-29519865-w9n5z\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.328927 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fed7859-0158-469a-bec9-275064e23cf6-config-volume\") pod \"collect-profiles-29519865-w9n5z\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.337045 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fed7859-0158-469a-bec9-275064e23cf6-secret-volume\") pod \"collect-profiles-29519865-w9n5z\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.344609 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkr4n\" (UniqueName: \"kubernetes.io/projected/5fed7859-0158-469a-bec9-275064e23cf6-kube-api-access-dkr4n\") pod \"collect-profiles-29519865-w9n5z\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.485171 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:00 crc kubenswrapper[4735]: I0215 21:45:00.937606 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z"] Feb 15 21:45:01 crc kubenswrapper[4735]: I0215 21:45:01.214095 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" event={"ID":"5fed7859-0158-469a-bec9-275064e23cf6","Type":"ContainerStarted","Data":"49ec5286198a90136cca12cf6199d69fc3f59c97ad2b7cb3f912d5cae06356fe"} Feb 15 21:45:01 crc kubenswrapper[4735]: I0215 21:45:01.214404 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" event={"ID":"5fed7859-0158-469a-bec9-275064e23cf6","Type":"ContainerStarted","Data":"b601e0662d31f0cffd1f8c86a7e6f851f31749ec0b20736f980d11393cba064b"} Feb 15 21:45:01 crc kubenswrapper[4735]: I0215 21:45:01.233173 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" podStartSLOduration=1.23315607 podStartE2EDuration="1.23315607s" podCreationTimestamp="2026-02-15 21:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-15 21:45:01.227604721 +0000 UTC m=+5309.093620344" watchObservedRunningTime="2026-02-15 21:45:01.23315607 +0000 UTC m=+5309.099171693" Feb 15 21:45:02 crc kubenswrapper[4735]: I0215 21:45:02.231214 4735 generic.go:334] "Generic (PLEG): container finished" podID="5fed7859-0158-469a-bec9-275064e23cf6" containerID="49ec5286198a90136cca12cf6199d69fc3f59c97ad2b7cb3f912d5cae06356fe" exitCode=0 Feb 15 21:45:02 crc kubenswrapper[4735]: I0215 21:45:02.231281 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" event={"ID":"5fed7859-0158-469a-bec9-275064e23cf6","Type":"ContainerDied","Data":"49ec5286198a90136cca12cf6199d69fc3f59c97ad2b7cb3f912d5cae06356fe"} Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.560932 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.697413 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fed7859-0158-469a-bec9-275064e23cf6-config-volume\") pod \"5fed7859-0158-469a-bec9-275064e23cf6\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.697484 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fed7859-0158-469a-bec9-275064e23cf6-secret-volume\") pod \"5fed7859-0158-469a-bec9-275064e23cf6\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.697524 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkr4n\" (UniqueName: \"kubernetes.io/projected/5fed7859-0158-469a-bec9-275064e23cf6-kube-api-access-dkr4n\") pod \"5fed7859-0158-469a-bec9-275064e23cf6\" (UID: \"5fed7859-0158-469a-bec9-275064e23cf6\") " Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.698276 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fed7859-0158-469a-bec9-275064e23cf6-config-volume" (OuterVolumeSpecName: "config-volume") pod "5fed7859-0158-469a-bec9-275064e23cf6" (UID: "5fed7859-0158-469a-bec9-275064e23cf6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.703439 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fed7859-0158-469a-bec9-275064e23cf6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5fed7859-0158-469a-bec9-275064e23cf6" (UID: "5fed7859-0158-469a-bec9-275064e23cf6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.703485 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fed7859-0158-469a-bec9-275064e23cf6-kube-api-access-dkr4n" (OuterVolumeSpecName: "kube-api-access-dkr4n") pod "5fed7859-0158-469a-bec9-275064e23cf6" (UID: "5fed7859-0158-469a-bec9-275064e23cf6"). InnerVolumeSpecName "kube-api-access-dkr4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.799997 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fed7859-0158-469a-bec9-275064e23cf6-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.800300 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkr4n\" (UniqueName: \"kubernetes.io/projected/5fed7859-0158-469a-bec9-275064e23cf6-kube-api-access-dkr4n\") on node \"crc\" DevicePath \"\"" Feb 15 21:45:03 crc kubenswrapper[4735]: I0215 21:45:03.800357 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fed7859-0158-469a-bec9-275064e23cf6-config-volume\") on node \"crc\" DevicePath \"\"" Feb 15 21:45:04 crc kubenswrapper[4735]: I0215 21:45:04.249853 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" event={"ID":"5fed7859-0158-469a-bec9-275064e23cf6","Type":"ContainerDied","Data":"b601e0662d31f0cffd1f8c86a7e6f851f31749ec0b20736f980d11393cba064b"} Feb 15 21:45:04 crc kubenswrapper[4735]: I0215 21:45:04.250189 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b601e0662d31f0cffd1f8c86a7e6f851f31749ec0b20736f980d11393cba064b" Feb 15 21:45:04 crc kubenswrapper[4735]: I0215 21:45:04.249892 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29519865-w9n5z" Feb 15 21:45:04 crc kubenswrapper[4735]: I0215 21:45:04.316416 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx"] Feb 15 21:45:04 crc kubenswrapper[4735]: I0215 21:45:04.324667 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29519820-kfshx"] Feb 15 21:45:04 crc kubenswrapper[4735]: I0215 21:45:04.906842 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16dfc773-d909-49eb-a20d-73ecf0a90271" path="/var/lib/kubelet/pods/16dfc773-d909-49eb-a20d-73ecf0a90271/volumes" Feb 15 21:45:07 crc kubenswrapper[4735]: I0215 21:45:07.887288 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:45:07 crc kubenswrapper[4735]: E0215 21:45:07.888091 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.389020 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8rzpj"] Feb 15 21:45:19 crc kubenswrapper[4735]: E0215 21:45:19.390086 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fed7859-0158-469a-bec9-275064e23cf6" containerName="collect-profiles" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.390104 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fed7859-0158-469a-bec9-275064e23cf6" containerName="collect-profiles" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.390348 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fed7859-0158-469a-bec9-275064e23cf6" containerName="collect-profiles" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.392061 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.408833 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8rzpj"] Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.544507 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-catalog-content\") pod \"redhat-operators-8rzpj\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.544605 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq65w\" (UniqueName: \"kubernetes.io/projected/76914201-10eb-4e49-91ef-68090547b386-kube-api-access-tq65w\") pod \"redhat-operators-8rzpj\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.544634 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-utilities\") pod \"redhat-operators-8rzpj\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.657720 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq65w\" (UniqueName: \"kubernetes.io/projected/76914201-10eb-4e49-91ef-68090547b386-kube-api-access-tq65w\") pod \"redhat-operators-8rzpj\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.657795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-utilities\") pod \"redhat-operators-8rzpj\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.657965 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-catalog-content\") pod \"redhat-operators-8rzpj\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.658524 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-utilities\") pod \"redhat-operators-8rzpj\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.658578 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-catalog-content\") pod \"redhat-operators-8rzpj\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.682898 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq65w\" (UniqueName: \"kubernetes.io/projected/76914201-10eb-4e49-91ef-68090547b386-kube-api-access-tq65w\") pod \"redhat-operators-8rzpj\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:19 crc kubenswrapper[4735]: I0215 21:45:19.715700 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:20 crc kubenswrapper[4735]: I0215 21:45:20.196842 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8rzpj"] Feb 15 21:45:20 crc kubenswrapper[4735]: I0215 21:45:20.392576 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rzpj" event={"ID":"76914201-10eb-4e49-91ef-68090547b386","Type":"ContainerStarted","Data":"576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a"} Feb 15 21:45:20 crc kubenswrapper[4735]: I0215 21:45:20.393742 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rzpj" event={"ID":"76914201-10eb-4e49-91ef-68090547b386","Type":"ContainerStarted","Data":"4895e49c35d836f1cdeffd6385e53f31ab2cd65b0972d9296b70aec550a28a22"} Feb 15 21:45:21 crc kubenswrapper[4735]: I0215 21:45:21.400638 4735 generic.go:334] "Generic (PLEG): container finished" podID="76914201-10eb-4e49-91ef-68090547b386" containerID="576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a" exitCode=0 Feb 15 21:45:21 crc kubenswrapper[4735]: I0215 21:45:21.400710 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rzpj" event={"ID":"76914201-10eb-4e49-91ef-68090547b386","Type":"ContainerDied","Data":"576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a"} Feb 15 21:45:22 crc kubenswrapper[4735]: I0215 21:45:22.414537 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rzpj" event={"ID":"76914201-10eb-4e49-91ef-68090547b386","Type":"ContainerStarted","Data":"cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca"} Feb 15 21:45:22 crc kubenswrapper[4735]: I0215 21:45:22.899899 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:45:22 crc kubenswrapper[4735]: E0215 21:45:22.900295 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:45:27 crc kubenswrapper[4735]: I0215 21:45:27.457334 4735 generic.go:334] "Generic (PLEG): container finished" podID="76914201-10eb-4e49-91ef-68090547b386" containerID="cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca" exitCode=0 Feb 15 21:45:27 crc kubenswrapper[4735]: I0215 21:45:27.457391 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rzpj" event={"ID":"76914201-10eb-4e49-91ef-68090547b386","Type":"ContainerDied","Data":"cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca"} Feb 15 21:45:28 crc kubenswrapper[4735]: I0215 21:45:28.468575 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rzpj" event={"ID":"76914201-10eb-4e49-91ef-68090547b386","Type":"ContainerStarted","Data":"894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c"} Feb 15 21:45:28 crc kubenswrapper[4735]: I0215 21:45:28.507773 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8rzpj" podStartSLOduration=2.909182286 podStartE2EDuration="9.507755336s" podCreationTimestamp="2026-02-15 21:45:19 +0000 UTC" firstStartedPulling="2026-02-15 21:45:21.403230598 +0000 UTC m=+5329.269246211" lastFinishedPulling="2026-02-15 21:45:28.001803628 +0000 UTC m=+5335.867819261" observedRunningTime="2026-02-15 21:45:28.496823563 +0000 UTC m=+5336.362839196" watchObservedRunningTime="2026-02-15 21:45:28.507755336 +0000 UTC m=+5336.373770959" Feb 15 21:45:29 crc kubenswrapper[4735]: I0215 21:45:29.716053 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:29 crc kubenswrapper[4735]: I0215 21:45:29.716371 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:30 crc kubenswrapper[4735]: I0215 21:45:30.770697 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8rzpj" podUID="76914201-10eb-4e49-91ef-68090547b386" containerName="registry-server" probeResult="failure" output=< Feb 15 21:45:30 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Feb 15 21:45:30 crc kubenswrapper[4735]: > Feb 15 21:45:35 crc kubenswrapper[4735]: I0215 21:45:35.887248 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:45:35 crc kubenswrapper[4735]: E0215 21:45:35.888435 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:45:39 crc kubenswrapper[4735]: I0215 21:45:39.773302 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:39 crc kubenswrapper[4735]: I0215 21:45:39.840862 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:40 crc kubenswrapper[4735]: I0215 21:45:40.024109 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8rzpj"] Feb 15 21:45:41 crc kubenswrapper[4735]: I0215 21:45:41.597759 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8rzpj" podUID="76914201-10eb-4e49-91ef-68090547b386" containerName="registry-server" containerID="cri-o://894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c" gracePeriod=2 Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.153121 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.329613 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq65w\" (UniqueName: \"kubernetes.io/projected/76914201-10eb-4e49-91ef-68090547b386-kube-api-access-tq65w\") pod \"76914201-10eb-4e49-91ef-68090547b386\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.329986 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-utilities\") pod \"76914201-10eb-4e49-91ef-68090547b386\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.330027 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-catalog-content\") pod \"76914201-10eb-4e49-91ef-68090547b386\" (UID: \"76914201-10eb-4e49-91ef-68090547b386\") " Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.330868 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-utilities" (OuterVolumeSpecName: "utilities") pod "76914201-10eb-4e49-91ef-68090547b386" (UID: "76914201-10eb-4e49-91ef-68090547b386"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.339202 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76914201-10eb-4e49-91ef-68090547b386-kube-api-access-tq65w" (OuterVolumeSpecName: "kube-api-access-tq65w") pod "76914201-10eb-4e49-91ef-68090547b386" (UID: "76914201-10eb-4e49-91ef-68090547b386"). InnerVolumeSpecName "kube-api-access-tq65w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.432270 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq65w\" (UniqueName: \"kubernetes.io/projected/76914201-10eb-4e49-91ef-68090547b386-kube-api-access-tq65w\") on node \"crc\" DevicePath \"\"" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.432306 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.455125 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76914201-10eb-4e49-91ef-68090547b386" (UID: "76914201-10eb-4e49-91ef-68090547b386"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.534304 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76914201-10eb-4e49-91ef-68090547b386-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.609896 4735 generic.go:334] "Generic (PLEG): container finished" podID="76914201-10eb-4e49-91ef-68090547b386" containerID="894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c" exitCode=0 Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.610008 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rzpj" event={"ID":"76914201-10eb-4e49-91ef-68090547b386","Type":"ContainerDied","Data":"894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c"} Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.610301 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rzpj" event={"ID":"76914201-10eb-4e49-91ef-68090547b386","Type":"ContainerDied","Data":"4895e49c35d836f1cdeffd6385e53f31ab2cd65b0972d9296b70aec550a28a22"} Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.610069 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8rzpj" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.610319 4735 scope.go:117] "RemoveContainer" containerID="894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.638708 4735 scope.go:117] "RemoveContainer" containerID="cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.654899 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8rzpj"] Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.667141 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8rzpj"] Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.681129 4735 scope.go:117] "RemoveContainer" containerID="576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.704163 4735 scope.go:117] "RemoveContainer" containerID="894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c" Feb 15 21:45:42 crc kubenswrapper[4735]: E0215 21:45:42.704676 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c\": container with ID starting with 894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c not found: ID does not exist" containerID="894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.704716 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c"} err="failed to get container status \"894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c\": rpc error: code = NotFound desc = could not find container \"894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c\": container with ID starting with 894670f4759ef161ce3fcbbb34735cd9bf053ec50fc0daf0afab0cd0f3e4270c not found: ID does not exist" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.704783 4735 scope.go:117] "RemoveContainer" containerID="cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca" Feb 15 21:45:42 crc kubenswrapper[4735]: E0215 21:45:42.705186 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca\": container with ID starting with cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca not found: ID does not exist" containerID="cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.705220 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca"} err="failed to get container status \"cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca\": rpc error: code = NotFound desc = could not find container \"cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca\": container with ID starting with cf1611b2008de9bcedff7f68ca3f569df5ebe1439babf85dc78d51de056503ca not found: ID does not exist" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.705234 4735 scope.go:117] "RemoveContainer" containerID="576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a" Feb 15 21:45:42 crc kubenswrapper[4735]: E0215 21:45:42.705476 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a\": container with ID starting with 576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a not found: ID does not exist" containerID="576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.705507 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a"} err="failed to get container status \"576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a\": rpc error: code = NotFound desc = could not find container \"576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a\": container with ID starting with 576d558d0fb32de3117e69edbfdec62e0478745b084d582e2af6e3530f52a31a not found: ID does not exist" Feb 15 21:45:42 crc kubenswrapper[4735]: I0215 21:45:42.901628 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76914201-10eb-4e49-91ef-68090547b386" path="/var/lib/kubelet/pods/76914201-10eb-4e49-91ef-68090547b386/volumes" Feb 15 21:45:46 crc kubenswrapper[4735]: I0215 21:45:46.887022 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:45:46 crc kubenswrapper[4735]: E0215 21:45:46.888249 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:45:54 crc kubenswrapper[4735]: I0215 21:45:54.217256 4735 scope.go:117] "RemoveContainer" containerID="461dc2b8e559b02c9d6e7ac3a52c974d8506c855c040c573c864b6a915de18cd" Feb 15 21:45:59 crc kubenswrapper[4735]: I0215 21:45:59.887365 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:45:59 crc kubenswrapper[4735]: E0215 21:45:59.888532 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:46:11 crc kubenswrapper[4735]: I0215 21:46:11.886940 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:46:11 crc kubenswrapper[4735]: E0215 21:46:11.887853 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.356022 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m45bb"] Feb 15 21:46:22 crc kubenswrapper[4735]: E0215 21:46:22.356889 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76914201-10eb-4e49-91ef-68090547b386" containerName="registry-server" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.356902 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="76914201-10eb-4e49-91ef-68090547b386" containerName="registry-server" Feb 15 21:46:22 crc kubenswrapper[4735]: E0215 21:46:22.356917 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76914201-10eb-4e49-91ef-68090547b386" containerName="extract-utilities" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.356923 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="76914201-10eb-4e49-91ef-68090547b386" containerName="extract-utilities" Feb 15 21:46:22 crc kubenswrapper[4735]: E0215 21:46:22.356935 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76914201-10eb-4e49-91ef-68090547b386" containerName="extract-content" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.356955 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="76914201-10eb-4e49-91ef-68090547b386" containerName="extract-content" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.357135 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="76914201-10eb-4e49-91ef-68090547b386" containerName="registry-server" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.358357 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.376618 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m45bb"] Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.400665 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-utilities\") pod \"community-operators-m45bb\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.400721 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x4qc\" (UniqueName: \"kubernetes.io/projected/1afa0259-15d2-4268-a007-704d0898239b-kube-api-access-6x4qc\") pod \"community-operators-m45bb\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.400760 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-catalog-content\") pod \"community-operators-m45bb\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.502728 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-catalog-content\") pod \"community-operators-m45bb\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.502966 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-utilities\") pod \"community-operators-m45bb\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.503026 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x4qc\" (UniqueName: \"kubernetes.io/projected/1afa0259-15d2-4268-a007-704d0898239b-kube-api-access-6x4qc\") pod \"community-operators-m45bb\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.503247 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-catalog-content\") pod \"community-operators-m45bb\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.503438 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-utilities\") pod \"community-operators-m45bb\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.525463 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x4qc\" (UniqueName: \"kubernetes.io/projected/1afa0259-15d2-4268-a007-704d0898239b-kube-api-access-6x4qc\") pod \"community-operators-m45bb\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:22 crc kubenswrapper[4735]: I0215 21:46:22.716319 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:23 crc kubenswrapper[4735]: I0215 21:46:23.111527 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m45bb"] Feb 15 21:46:24 crc kubenswrapper[4735]: I0215 21:46:24.029854 4735 generic.go:334] "Generic (PLEG): container finished" podID="1afa0259-15d2-4268-a007-704d0898239b" containerID="c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d" exitCode=0 Feb 15 21:46:24 crc kubenswrapper[4735]: I0215 21:46:24.029961 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m45bb" event={"ID":"1afa0259-15d2-4268-a007-704d0898239b","Type":"ContainerDied","Data":"c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d"} Feb 15 21:46:24 crc kubenswrapper[4735]: I0215 21:46:24.032076 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m45bb" event={"ID":"1afa0259-15d2-4268-a007-704d0898239b","Type":"ContainerStarted","Data":"b6a100e77f03e042ade1cb8e0d76b4cd0204e05d4106259d95259ce0afe87a93"} Feb 15 21:46:25 crc kubenswrapper[4735]: I0215 21:46:25.050398 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m45bb" event={"ID":"1afa0259-15d2-4268-a007-704d0898239b","Type":"ContainerStarted","Data":"357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249"} Feb 15 21:46:25 crc kubenswrapper[4735]: I0215 21:46:25.886733 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:46:25 crc kubenswrapper[4735]: E0215 21:46:25.887169 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:46:27 crc kubenswrapper[4735]: I0215 21:46:27.077484 4735 generic.go:334] "Generic (PLEG): container finished" podID="1afa0259-15d2-4268-a007-704d0898239b" containerID="357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249" exitCode=0 Feb 15 21:46:27 crc kubenswrapper[4735]: I0215 21:46:27.077537 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m45bb" event={"ID":"1afa0259-15d2-4268-a007-704d0898239b","Type":"ContainerDied","Data":"357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249"} Feb 15 21:46:28 crc kubenswrapper[4735]: I0215 21:46:28.107822 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m45bb" event={"ID":"1afa0259-15d2-4268-a007-704d0898239b","Type":"ContainerStarted","Data":"c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3"} Feb 15 21:46:28 crc kubenswrapper[4735]: I0215 21:46:28.139462 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m45bb" podStartSLOduration=2.7210138600000002 podStartE2EDuration="6.139440077s" podCreationTimestamp="2026-02-15 21:46:22 +0000 UTC" firstStartedPulling="2026-02-15 21:46:24.032340898 +0000 UTC m=+5391.898356561" lastFinishedPulling="2026-02-15 21:46:27.450767105 +0000 UTC m=+5395.316782778" observedRunningTime="2026-02-15 21:46:28.135094821 +0000 UTC m=+5396.001110454" watchObservedRunningTime="2026-02-15 21:46:28.139440077 +0000 UTC m=+5396.005455710" Feb 15 21:46:32 crc kubenswrapper[4735]: I0215 21:46:32.716519 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:32 crc kubenswrapper[4735]: I0215 21:46:32.717124 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:32 crc kubenswrapper[4735]: I0215 21:46:32.765297 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:33 crc kubenswrapper[4735]: I0215 21:46:33.200822 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:33 crc kubenswrapper[4735]: I0215 21:46:33.348521 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m45bb"] Feb 15 21:46:35 crc kubenswrapper[4735]: I0215 21:46:35.172362 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m45bb" podUID="1afa0259-15d2-4268-a007-704d0898239b" containerName="registry-server" containerID="cri-o://c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3" gracePeriod=2 Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.166021 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.184602 4735 generic.go:334] "Generic (PLEG): container finished" podID="1afa0259-15d2-4268-a007-704d0898239b" containerID="c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3" exitCode=0 Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.184662 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m45bb" event={"ID":"1afa0259-15d2-4268-a007-704d0898239b","Type":"ContainerDied","Data":"c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3"} Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.184697 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m45bb" event={"ID":"1afa0259-15d2-4268-a007-704d0898239b","Type":"ContainerDied","Data":"b6a100e77f03e042ade1cb8e0d76b4cd0204e05d4106259d95259ce0afe87a93"} Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.184715 4735 scope.go:117] "RemoveContainer" containerID="c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.184848 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m45bb" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.255055 4735 scope.go:117] "RemoveContainer" containerID="357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.279074 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-catalog-content\") pod \"1afa0259-15d2-4268-a007-704d0898239b\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.279176 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-utilities\") pod \"1afa0259-15d2-4268-a007-704d0898239b\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.279268 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x4qc\" (UniqueName: \"kubernetes.io/projected/1afa0259-15d2-4268-a007-704d0898239b-kube-api-access-6x4qc\") pod \"1afa0259-15d2-4268-a007-704d0898239b\" (UID: \"1afa0259-15d2-4268-a007-704d0898239b\") " Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.280142 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-utilities" (OuterVolumeSpecName: "utilities") pod "1afa0259-15d2-4268-a007-704d0898239b" (UID: "1afa0259-15d2-4268-a007-704d0898239b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.284071 4735 scope.go:117] "RemoveContainer" containerID="c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.287430 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1afa0259-15d2-4268-a007-704d0898239b-kube-api-access-6x4qc" (OuterVolumeSpecName: "kube-api-access-6x4qc") pod "1afa0259-15d2-4268-a007-704d0898239b" (UID: "1afa0259-15d2-4268-a007-704d0898239b"). InnerVolumeSpecName "kube-api-access-6x4qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.340897 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1afa0259-15d2-4268-a007-704d0898239b" (UID: "1afa0259-15d2-4268-a007-704d0898239b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.366194 4735 scope.go:117] "RemoveContainer" containerID="c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3" Feb 15 21:46:36 crc kubenswrapper[4735]: E0215 21:46:36.367036 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3\": container with ID starting with c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3 not found: ID does not exist" containerID="c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.367074 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3"} err="failed to get container status \"c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3\": rpc error: code = NotFound desc = could not find container \"c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3\": container with ID starting with c864de1b0e2e5aab99c0cfb2d3916712eff7bf353212b07d70f0a7ad801f37e3 not found: ID does not exist" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.367100 4735 scope.go:117] "RemoveContainer" containerID="357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249" Feb 15 21:46:36 crc kubenswrapper[4735]: E0215 21:46:36.367460 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249\": container with ID starting with 357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249 not found: ID does not exist" containerID="357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.367490 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249"} err="failed to get container status \"357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249\": rpc error: code = NotFound desc = could not find container \"357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249\": container with ID starting with 357e8e4a5ed2359fa216c218d2a65fb559f2c455c6a3751cd982877ce49d0249 not found: ID does not exist" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.367512 4735 scope.go:117] "RemoveContainer" containerID="c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d" Feb 15 21:46:36 crc kubenswrapper[4735]: E0215 21:46:36.367839 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d\": container with ID starting with c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d not found: ID does not exist" containerID="c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.367869 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d"} err="failed to get container status \"c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d\": rpc error: code = NotFound desc = could not find container \"c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d\": container with ID starting with c5deff5200f42c541bc4772648f809b8e445cb08a694bdc432417df53d4f373d not found: ID does not exist" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.380627 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x4qc\" (UniqueName: \"kubernetes.io/projected/1afa0259-15d2-4268-a007-704d0898239b-kube-api-access-6x4qc\") on node \"crc\" DevicePath \"\"" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.380655 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.380667 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afa0259-15d2-4268-a007-704d0898239b-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.527754 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m45bb"] Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.536552 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m45bb"] Feb 15 21:46:36 crc kubenswrapper[4735]: I0215 21:46:36.912174 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1afa0259-15d2-4268-a007-704d0898239b" path="/var/lib/kubelet/pods/1afa0259-15d2-4268-a007-704d0898239b/volumes" Feb 15 21:46:39 crc kubenswrapper[4735]: I0215 21:46:39.887642 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:46:39 crc kubenswrapper[4735]: E0215 21:46:39.888535 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:46:51 crc kubenswrapper[4735]: I0215 21:46:51.887370 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:46:51 crc kubenswrapper[4735]: E0215 21:46:51.888527 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:47:06 crc kubenswrapper[4735]: I0215 21:47:06.887045 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:47:06 crc kubenswrapper[4735]: E0215 21:47:06.888398 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:47:20 crc kubenswrapper[4735]: I0215 21:47:20.887043 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:47:20 crc kubenswrapper[4735]: E0215 21:47:20.887993 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:47:34 crc kubenswrapper[4735]: I0215 21:47:34.887129 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:47:34 crc kubenswrapper[4735]: E0215 21:47:34.887960 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kjps6_openshift-machine-config-operator(43de4639-a922-4182-9377-d1f28b3aa4c1)\"" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" podUID="43de4639-a922-4182-9377-d1f28b3aa4c1" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.360173 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hk4md"] Feb 15 21:47:48 crc kubenswrapper[4735]: E0215 21:47:48.361039 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afa0259-15d2-4268-a007-704d0898239b" containerName="extract-content" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.361054 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afa0259-15d2-4268-a007-704d0898239b" containerName="extract-content" Feb 15 21:47:48 crc kubenswrapper[4735]: E0215 21:47:48.361072 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afa0259-15d2-4268-a007-704d0898239b" containerName="registry-server" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.361080 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afa0259-15d2-4268-a007-704d0898239b" containerName="registry-server" Feb 15 21:47:48 crc kubenswrapper[4735]: E0215 21:47:48.361115 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afa0259-15d2-4268-a007-704d0898239b" containerName="extract-utilities" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.361123 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afa0259-15d2-4268-a007-704d0898239b" containerName="extract-utilities" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.361368 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1afa0259-15d2-4268-a007-704d0898239b" containerName="registry-server" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.363022 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.378078 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hk4md"] Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.409036 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-catalog-content\") pod \"redhat-marketplace-hk4md\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.409089 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-utilities\") pod \"redhat-marketplace-hk4md\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.409162 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9j77\" (UniqueName: \"kubernetes.io/projected/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-kube-api-access-h9j77\") pod \"redhat-marketplace-hk4md\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.510615 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9j77\" (UniqueName: \"kubernetes.io/projected/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-kube-api-access-h9j77\") pod \"redhat-marketplace-hk4md\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.511223 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-catalog-content\") pod \"redhat-marketplace-hk4md\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.511802 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-catalog-content\") pod \"redhat-marketplace-hk4md\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.511886 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-utilities\") pod \"redhat-marketplace-hk4md\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.512242 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-utilities\") pod \"redhat-marketplace-hk4md\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.528221 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9j77\" (UniqueName: \"kubernetes.io/projected/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-kube-api-access-h9j77\") pod \"redhat-marketplace-hk4md\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:48 crc kubenswrapper[4735]: I0215 21:47:48.686344 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:49 crc kubenswrapper[4735]: W0215 21:47:49.168004 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf93d6cf_c47f_4abb_9729_18ac6f1ec1ef.slice/crio-72b48ea17b6db374308666eecb0f2240999778291f9d09c7e9f606e33e1f81e8 WatchSource:0}: Error finding container 72b48ea17b6db374308666eecb0f2240999778291f9d09c7e9f606e33e1f81e8: Status 404 returned error can't find the container with id 72b48ea17b6db374308666eecb0f2240999778291f9d09c7e9f606e33e1f81e8 Feb 15 21:47:49 crc kubenswrapper[4735]: I0215 21:47:49.179724 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hk4md"] Feb 15 21:47:49 crc kubenswrapper[4735]: I0215 21:47:49.886565 4735 scope.go:117] "RemoveContainer" containerID="b990f87a3064ea7fafa4a4a7f48acb9e03b0161b9648198604ffe76c547ef89b" Feb 15 21:47:49 crc kubenswrapper[4735]: I0215 21:47:49.893576 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk4md" event={"ID":"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef","Type":"ContainerDied","Data":"e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8"} Feb 15 21:47:49 crc kubenswrapper[4735]: I0215 21:47:49.893471 4735 generic.go:334] "Generic (PLEG): container finished" podID="bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef" containerID="e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8" exitCode=0 Feb 15 21:47:49 crc kubenswrapper[4735]: I0215 21:47:49.893879 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk4md" event={"ID":"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef","Type":"ContainerStarted","Data":"72b48ea17b6db374308666eecb0f2240999778291f9d09c7e9f606e33e1f81e8"} Feb 15 21:47:50 crc kubenswrapper[4735]: I0215 21:47:50.940319 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kjps6" event={"ID":"43de4639-a922-4182-9377-d1f28b3aa4c1","Type":"ContainerStarted","Data":"3cde6c673eaf02d239f15ef6956f43b401fe09a8a060898ea58a7d8192054089"} Feb 15 21:47:51 crc kubenswrapper[4735]: I0215 21:47:51.952755 4735 generic.go:334] "Generic (PLEG): container finished" podID="bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef" containerID="230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558" exitCode=0 Feb 15 21:47:51 crc kubenswrapper[4735]: I0215 21:47:51.952826 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk4md" event={"ID":"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef","Type":"ContainerDied","Data":"230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558"} Feb 15 21:47:52 crc kubenswrapper[4735]: I0215 21:47:52.963207 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk4md" event={"ID":"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef","Type":"ContainerStarted","Data":"e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877"} Feb 15 21:47:52 crc kubenswrapper[4735]: I0215 21:47:52.990937 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hk4md" podStartSLOduration=2.524163571 podStartE2EDuration="4.99091328s" podCreationTimestamp="2026-02-15 21:47:48 +0000 UTC" firstStartedPulling="2026-02-15 21:47:49.902847948 +0000 UTC m=+5477.768863571" lastFinishedPulling="2026-02-15 21:47:52.369597657 +0000 UTC m=+5480.235613280" observedRunningTime="2026-02-15 21:47:52.98197687 +0000 UTC m=+5480.847992503" watchObservedRunningTime="2026-02-15 21:47:52.99091328 +0000 UTC m=+5480.856928903" Feb 15 21:47:58 crc kubenswrapper[4735]: I0215 21:47:58.686424 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:58 crc kubenswrapper[4735]: I0215 21:47:58.686971 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:58 crc kubenswrapper[4735]: I0215 21:47:58.735089 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:59 crc kubenswrapper[4735]: I0215 21:47:59.066802 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:47:59 crc kubenswrapper[4735]: I0215 21:47:59.121673 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hk4md"] Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.031384 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hk4md" podUID="bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef" containerName="registry-server" containerID="cri-o://e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877" gracePeriod=2 Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.464842 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.656396 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-utilities\") pod \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.656632 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-catalog-content\") pod \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.656699 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9j77\" (UniqueName: \"kubernetes.io/projected/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-kube-api-access-h9j77\") pod \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\" (UID: \"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef\") " Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.657405 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-utilities" (OuterVolumeSpecName: "utilities") pod "bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef" (UID: "bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.669412 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-kube-api-access-h9j77" (OuterVolumeSpecName: "kube-api-access-h9j77") pod "bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef" (UID: "bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef"). InnerVolumeSpecName "kube-api-access-h9j77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.681624 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef" (UID: "bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.758899 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.758930 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9j77\" (UniqueName: \"kubernetes.io/projected/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-kube-api-access-h9j77\") on node \"crc\" DevicePath \"\"" Feb 15 21:48:01 crc kubenswrapper[4735]: I0215 21:48:01.758941 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef-utilities\") on node \"crc\" DevicePath \"\"" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.040921 4735 generic.go:334] "Generic (PLEG): container finished" podID="bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef" containerID="e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877" exitCode=0 Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.040985 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk4md" event={"ID":"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef","Type":"ContainerDied","Data":"e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877"} Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.041013 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk4md" event={"ID":"bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef","Type":"ContainerDied","Data":"72b48ea17b6db374308666eecb0f2240999778291f9d09c7e9f606e33e1f81e8"} Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.041032 4735 scope.go:117] "RemoveContainer" containerID="e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.041088 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hk4md" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.062198 4735 scope.go:117] "RemoveContainer" containerID="230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.092461 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hk4md"] Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.101609 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hk4md"] Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.103330 4735 scope.go:117] "RemoveContainer" containerID="e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.142632 4735 scope.go:117] "RemoveContainer" containerID="e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877" Feb 15 21:48:02 crc kubenswrapper[4735]: E0215 21:48:02.143097 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877\": container with ID starting with e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877 not found: ID does not exist" containerID="e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.143155 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877"} err="failed to get container status \"e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877\": rpc error: code = NotFound desc = could not find container \"e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877\": container with ID starting with e3d304b01cc601bf15a1ea47b74446aa619cba31cad59271297e8655a6fd3877 not found: ID does not exist" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.143181 4735 scope.go:117] "RemoveContainer" containerID="230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558" Feb 15 21:48:02 crc kubenswrapper[4735]: E0215 21:48:02.143524 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558\": container with ID starting with 230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558 not found: ID does not exist" containerID="230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.143585 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558"} err="failed to get container status \"230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558\": rpc error: code = NotFound desc = could not find container \"230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558\": container with ID starting with 230751a8da234fcfad7b718a4a68b96306f379c905468c2df4b2d236c96f4558 not found: ID does not exist" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.143619 4735 scope.go:117] "RemoveContainer" containerID="e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8" Feb 15 21:48:02 crc kubenswrapper[4735]: E0215 21:48:02.143900 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8\": container with ID starting with e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8 not found: ID does not exist" containerID="e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.143933 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8"} err="failed to get container status \"e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8\": rpc error: code = NotFound desc = could not find container \"e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8\": container with ID starting with e2387e248a86f0a45677c82141166c5745ddf23d12e4f04da16e8d2232e694c8 not found: ID does not exist" Feb 15 21:48:02 crc kubenswrapper[4735]: I0215 21:48:02.896530 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef" path="/var/lib/kubelet/pods/bf93d6cf-c47f-4abb-9729-18ac6f1ec1ef/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515144437600024451 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015144437601017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015144424235016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015144424235015461 5ustar corecore